var/home/core/zuul-output/0000755000175000017500000000000015112337072014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112343600015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005001221615112343571017674 0ustar rootrootNov 28 15:26:12 crc systemd[1]: Starting Kubernetes Kubelet... Nov 28 15:26:12 crc restorecon[4747]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 15:26:12 crc restorecon[4747]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 28 15:26:13 crc kubenswrapper[4805]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 15:26:13 crc kubenswrapper[4805]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 28 15:26:13 crc kubenswrapper[4805]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 15:26:13 crc kubenswrapper[4805]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 15:26:13 crc kubenswrapper[4805]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 28 15:26:13 crc kubenswrapper[4805]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.055106 4805 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060417 4805 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060470 4805 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060474 4805 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060479 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060484 4805 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060490 4805 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060497 4805 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060502 4805 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060507 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060512 4805 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060518 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060523 4805 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060528 4805 feature_gate.go:330] unrecognized feature gate: Example Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060533 4805 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060538 4805 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060544 4805 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060549 4805 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060554 4805 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060558 4805 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060564 4805 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060569 4805 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060577 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060582 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060587 4805 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060590 4805 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060594 4805 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060598 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060604 4805 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060609 4805 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060613 4805 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060617 4805 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060621 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060625 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060638 4805 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060644 4805 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060649 4805 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060654 4805 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060659 4805 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060666 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060671 4805 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060675 4805 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060681 4805 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060687 4805 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060692 4805 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060695 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060699 4805 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060703 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060706 4805 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060710 4805 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060715 4805 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060718 4805 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060722 4805 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060726 4805 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060733 4805 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060739 4805 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060743 4805 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060746 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060750 4805 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060754 4805 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060758 4805 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060761 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060764 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060770 4805 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060774 4805 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060780 4805 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060784 4805 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060788 4805 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060791 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060795 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060799 4805 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.060803 4805 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061060 4805 flags.go:64] FLAG: --address="0.0.0.0" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061079 4805 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061090 4805 flags.go:64] FLAG: --anonymous-auth="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061097 4805 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061105 4805 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061110 4805 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061123 4805 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061135 4805 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061141 4805 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061146 4805 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061152 4805 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061161 4805 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061167 4805 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061173 4805 flags.go:64] FLAG: --cgroup-root="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061180 4805 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061186 4805 flags.go:64] FLAG: --client-ca-file="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061191 4805 flags.go:64] FLAG: --cloud-config="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061196 4805 flags.go:64] FLAG: --cloud-provider="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061202 4805 flags.go:64] FLAG: --cluster-dns="[]" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061209 4805 flags.go:64] FLAG: --cluster-domain="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061215 4805 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061220 4805 flags.go:64] FLAG: --config-dir="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061225 4805 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061231 4805 flags.go:64] FLAG: --container-log-max-files="5" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061240 4805 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061245 4805 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061251 4805 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061256 4805 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061262 4805 flags.go:64] FLAG: --contention-profiling="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061268 4805 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061274 4805 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061279 4805 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061284 4805 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061292 4805 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061297 4805 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061303 4805 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061307 4805 flags.go:64] FLAG: --enable-load-reader="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061312 4805 flags.go:64] FLAG: --enable-server="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061317 4805 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061324 4805 flags.go:64] FLAG: --event-burst="100" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061329 4805 flags.go:64] FLAG: --event-qps="50" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061334 4805 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061339 4805 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061348 4805 flags.go:64] FLAG: --eviction-hard="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061384 4805 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061390 4805 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061395 4805 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061404 4805 flags.go:64] FLAG: --eviction-soft="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061409 4805 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061413 4805 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061418 4805 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061422 4805 flags.go:64] FLAG: --experimental-mounter-path="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061427 4805 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061432 4805 flags.go:64] FLAG: --fail-swap-on="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061436 4805 flags.go:64] FLAG: --feature-gates="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061444 4805 flags.go:64] FLAG: --file-check-frequency="20s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061452 4805 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061464 4805 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061471 4805 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061478 4805 flags.go:64] FLAG: --healthz-port="10248" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061483 4805 flags.go:64] FLAG: --help="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061488 4805 flags.go:64] FLAG: --hostname-override="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061492 4805 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061497 4805 flags.go:64] FLAG: --http-check-frequency="20s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061501 4805 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061506 4805 flags.go:64] FLAG: --image-credential-provider-config="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061513 4805 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061523 4805 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061531 4805 flags.go:64] FLAG: --image-service-endpoint="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061537 4805 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061543 4805 flags.go:64] FLAG: --kube-api-burst="100" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061548 4805 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061554 4805 flags.go:64] FLAG: --kube-api-qps="50" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061559 4805 flags.go:64] FLAG: --kube-reserved="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061564 4805 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061569 4805 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061575 4805 flags.go:64] FLAG: --kubelet-cgroups="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061581 4805 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061587 4805 flags.go:64] FLAG: --lock-file="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061594 4805 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061600 4805 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061606 4805 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061617 4805 flags.go:64] FLAG: --log-json-split-stream="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061629 4805 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061637 4805 flags.go:64] FLAG: --log-text-split-stream="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061643 4805 flags.go:64] FLAG: --logging-format="text" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061649 4805 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061655 4805 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061660 4805 flags.go:64] FLAG: --manifest-url="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061666 4805 flags.go:64] FLAG: --manifest-url-header="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061676 4805 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061681 4805 flags.go:64] FLAG: --max-open-files="1000000" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061688 4805 flags.go:64] FLAG: --max-pods="110" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061692 4805 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061697 4805 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061703 4805 flags.go:64] FLAG: --memory-manager-policy="None" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061709 4805 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061715 4805 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061720 4805 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061726 4805 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061740 4805 flags.go:64] FLAG: --node-status-max-images="50" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061745 4805 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061751 4805 flags.go:64] FLAG: --oom-score-adj="-999" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061756 4805 flags.go:64] FLAG: --pod-cidr="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061761 4805 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061772 4805 flags.go:64] FLAG: --pod-manifest-path="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061778 4805 flags.go:64] FLAG: --pod-max-pids="-1" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061783 4805 flags.go:64] FLAG: --pods-per-core="0" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061788 4805 flags.go:64] FLAG: --port="10250" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061793 4805 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061799 4805 flags.go:64] FLAG: --provider-id="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061806 4805 flags.go:64] FLAG: --qos-reserved="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061812 4805 flags.go:64] FLAG: --read-only-port="10255" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061818 4805 flags.go:64] FLAG: --register-node="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061824 4805 flags.go:64] FLAG: --register-schedulable="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061829 4805 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061841 4805 flags.go:64] FLAG: --registry-burst="10" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061847 4805 flags.go:64] FLAG: --registry-qps="5" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061852 4805 flags.go:64] FLAG: --reserved-cpus="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061864 4805 flags.go:64] FLAG: --reserved-memory="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061873 4805 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061879 4805 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061885 4805 flags.go:64] FLAG: --rotate-certificates="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061891 4805 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061897 4805 flags.go:64] FLAG: --runonce="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.061903 4805 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062021 4805 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062071 4805 flags.go:64] FLAG: --seccomp-default="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062078 4805 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062084 4805 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062093 4805 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062100 4805 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062106 4805 flags.go:64] FLAG: --storage-driver-password="root" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062111 4805 flags.go:64] FLAG: --storage-driver-secure="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062116 4805 flags.go:64] FLAG: --storage-driver-table="stats" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062121 4805 flags.go:64] FLAG: --storage-driver-user="root" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062125 4805 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062131 4805 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062136 4805 flags.go:64] FLAG: --system-cgroups="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062143 4805 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062157 4805 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062161 4805 flags.go:64] FLAG: --tls-cert-file="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062166 4805 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062179 4805 flags.go:64] FLAG: --tls-min-version="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062184 4805 flags.go:64] FLAG: --tls-private-key-file="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062188 4805 flags.go:64] FLAG: --topology-manager-policy="none" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062192 4805 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062197 4805 flags.go:64] FLAG: --topology-manager-scope="container" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062201 4805 flags.go:64] FLAG: --v="2" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062241 4805 flags.go:64] FLAG: --version="false" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062248 4805 flags.go:64] FLAG: --vmodule="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062254 4805 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062259 4805 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062469 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062475 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062484 4805 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062488 4805 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062492 4805 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062496 4805 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062501 4805 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062504 4805 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062508 4805 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062512 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062516 4805 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062520 4805 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062524 4805 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062528 4805 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062532 4805 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062536 4805 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062540 4805 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062544 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062547 4805 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062551 4805 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062554 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062558 4805 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062571 4805 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062574 4805 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062579 4805 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062582 4805 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062621 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062626 4805 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062630 4805 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062634 4805 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062638 4805 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062642 4805 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062647 4805 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062650 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062654 4805 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062657 4805 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062662 4805 feature_gate.go:330] unrecognized feature gate: Example Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062665 4805 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062669 4805 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062672 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062676 4805 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062679 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062683 4805 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062687 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062690 4805 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062695 4805 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062700 4805 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062705 4805 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062709 4805 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062713 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062718 4805 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062722 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062726 4805 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062731 4805 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062737 4805 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062742 4805 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062749 4805 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062755 4805 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062761 4805 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062766 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062771 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062776 4805 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062780 4805 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062785 4805 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062789 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062793 4805 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062798 4805 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062803 4805 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062807 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062811 4805 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.062815 4805 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.062823 4805 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.073468 4805 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.073520 4805 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073684 4805 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073697 4805 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073707 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073718 4805 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073731 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073740 4805 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073749 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073758 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073766 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073774 4805 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073782 4805 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073789 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073797 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073805 4805 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073812 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073821 4805 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073828 4805 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073836 4805 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073844 4805 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073852 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073860 4805 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073867 4805 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073875 4805 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073883 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073893 4805 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073903 4805 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073911 4805 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073920 4805 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073929 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073936 4805 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073944 4805 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073955 4805 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073963 4805 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073971 4805 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073980 4805 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073988 4805 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.073996 4805 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074003 4805 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074011 4805 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074019 4805 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074027 4805 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074034 4805 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074043 4805 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074051 4805 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074059 4805 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074069 4805 feature_gate.go:330] unrecognized feature gate: Example Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074077 4805 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074088 4805 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074098 4805 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074107 4805 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074116 4805 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074124 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074132 4805 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074140 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074149 4805 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074157 4805 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074165 4805 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074173 4805 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074182 4805 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074190 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074200 4805 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074210 4805 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074218 4805 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074227 4805 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074235 4805 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074244 4805 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074252 4805 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074260 4805 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074268 4805 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074276 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074286 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.074298 4805 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074606 4805 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074623 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074633 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074643 4805 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074653 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074662 4805 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074671 4805 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074679 4805 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074687 4805 feature_gate.go:330] unrecognized feature gate: Example Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074695 4805 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074703 4805 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074711 4805 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074720 4805 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074729 4805 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074736 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074745 4805 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074752 4805 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074760 4805 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074769 4805 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074776 4805 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074784 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074793 4805 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074800 4805 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074808 4805 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074817 4805 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074825 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074832 4805 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074840 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074847 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074856 4805 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074863 4805 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074871 4805 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074879 4805 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074887 4805 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074896 4805 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074904 4805 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074912 4805 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074922 4805 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074932 4805 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074941 4805 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074950 4805 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074958 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074965 4805 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074973 4805 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074982 4805 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.074992 4805 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075002 4805 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075011 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075020 4805 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075029 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075037 4805 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075046 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075054 4805 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075064 4805 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075073 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075083 4805 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075091 4805 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075100 4805 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075108 4805 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075118 4805 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075127 4805 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075136 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075144 4805 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075152 4805 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075160 4805 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075169 4805 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075176 4805 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075185 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075192 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075200 4805 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.075209 4805 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.075222 4805 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.075534 4805 server.go:940] "Client rotation is on, will bootstrap in background" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.079919 4805 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.080074 4805 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.081025 4805 server.go:997] "Starting client certificate rotation" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.081068 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.081556 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-08 05:14:16.963864243 +0000 UTC Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.081704 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 973h48m3.882167269s for next certificate rotation Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.089069 4805 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.092123 4805 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.098906 4805 log.go:25] "Validated CRI v1 runtime API" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.112017 4805 log.go:25] "Validated CRI v1 image API" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.113899 4805 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.116613 4805 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-28-15-21-44-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.116660 4805 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.133176 4805 manager.go:217] Machine: {Timestamp:2025-11-28 15:26:13.131727922 +0000 UTC m=+0.181519253 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:164ad39d-af70-4732-b323-1a0ca23607cd BootID:b34b13c7-e3bf-4272-b34c-e52b7c30974a Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:73:6e:98 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:73:6e:98 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:bd:af:82 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:89:3d:07 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:92:4b:63 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:c0:64:38 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:18:fc:63 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:52:95:21:85:b1:e7 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a6:a3:db:3e:af:af Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.133478 4805 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.133655 4805 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.134140 4805 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.134534 4805 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.134576 4805 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.134811 4805 topology_manager.go:138] "Creating topology manager with none policy" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.134821 4805 container_manager_linux.go:303] "Creating device plugin manager" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.135000 4805 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.135040 4805 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.135245 4805 state_mem.go:36] "Initialized new in-memory state store" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.135534 4805 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.136131 4805 kubelet.go:418] "Attempting to sync node with API server" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.136154 4805 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.136180 4805 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.136195 4805 kubelet.go:324] "Adding apiserver pod source" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.136209 4805 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.137938 4805 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.138272 4805 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.138956 4805 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139517 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139544 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139553 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139562 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139576 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139631 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139644 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139658 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139667 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139675 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139708 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139719 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.139981 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.140185 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.140199 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.140336 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.196:6443: connect: connection refused" logger="UnhandledError" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.140349 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.196:6443: connect: connection refused" logger="UnhandledError" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.140507 4805 server.go:1280] "Started kubelet" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.140668 4805 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.140939 4805 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.140942 4805 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.141504 4805 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 28 15:26:13 crc systemd[1]: Started Kubernetes Kubelet. Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.143503 4805 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.143544 4805 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.143762 4805 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 04:36:26.519059331 +0000 UTC Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.143811 4805 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1213h10m13.375250969s for next certificate rotation Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.143851 4805 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.143936 4805 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.143968 4805 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.144064 4805 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.144268 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="200ms" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.145117 4805 server.go:460] "Adding debug handlers to kubelet server" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.143742 4805 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.196:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187c3521c2c7d955 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 15:26:13.140461909 +0000 UTC m=+0.190253240,LastTimestamp:2025-11-28 15:26:13.140461909 +0000 UTC m=+0.190253240,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.147934 4805 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.147974 4805 factory.go:55] Registering systemd factory Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.147988 4805 factory.go:221] Registration of the systemd container factory successfully Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.145144 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.149010 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.196:6443: connect: connection refused" logger="UnhandledError" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.152059 4805 factory.go:153] Registering CRI-O factory Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.152205 4805 factory.go:221] Registration of the crio container factory successfully Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.152313 4805 factory.go:103] Registering Raw factory Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.154589 4805 manager.go:1196] Started watching for new ooms in manager Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.155461 4805 manager.go:319] Starting recovery of all containers Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159350 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159462 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159478 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159489 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159503 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159514 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159525 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159536 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159550 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159569 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159583 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159598 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159612 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159624 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159637 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159648 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159659 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159673 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159686 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159698 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159711 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159723 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159734 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159747 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159759 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159808 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159827 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159843 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159891 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159907 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159922 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159946 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159961 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159980 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.159991 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160006 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160019 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160034 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160050 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160067 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160081 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160095 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160106 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160123 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160141 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160157 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160171 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160186 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160247 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160260 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160270 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160281 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160297 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160310 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160322 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160336 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160347 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160384 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160397 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160411 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160423 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160433 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160443 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160452 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160463 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160473 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160484 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160494 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160504 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160515 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160525 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160537 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160548 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160559 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160569 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160582 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160595 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160608 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160619 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160629 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160641 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160652 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160661 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160672 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160683 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160693 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160705 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160715 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160725 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160739 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160750 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160760 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160773 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160783 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160793 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160804 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160814 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160824 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160834 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160845 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160855 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160865 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160875 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160886 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160906 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160918 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160929 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160940 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160952 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160964 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160974 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160986 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.160997 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161009 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161018 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161028 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161039 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161050 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161060 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161071 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161080 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161097 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161107 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161121 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161134 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161773 4805 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161796 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.161807 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162341 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162385 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162399 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162412 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162425 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162435 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162447 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162458 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162475 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162490 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162501 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162511 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162526 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162537 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162548 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162560 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162570 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162583 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162593 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162606 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162616 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162628 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162638 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162649 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162659 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162670 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162681 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162693 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162708 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162722 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162735 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162747 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162760 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162789 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162799 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162809 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162820 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162832 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162843 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162854 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162865 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162877 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162888 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162898 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162912 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162923 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162933 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162944 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162956 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162967 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162978 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.162991 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163002 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163012 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163025 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163036 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163046 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163059 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163070 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163080 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163092 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163102 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163120 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163129 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163179 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163191 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163201 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163211 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163220 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163232 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163243 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163310 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163321 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163339 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163349 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163377 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163391 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163406 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163417 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163430 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163442 4805 reconstruct.go:97] "Volume reconstruction finished" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.163453 4805 reconciler.go:26] "Reconciler: start to sync state" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.175884 4805 manager.go:324] Recovery completed Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.189618 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.193729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.193797 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.193815 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.196062 4805 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.196090 4805 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.196112 4805 state_mem.go:36] "Initialized new in-memory state store" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.201101 4805 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.202644 4805 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.202682 4805 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.202703 4805 kubelet.go:2335] "Starting kubelet main sync loop" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.202743 4805 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.203696 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.203777 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.196:6443: connect: connection refused" logger="UnhandledError" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.205958 4805 policy_none.go:49] "None policy: Start" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.206586 4805 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.206616 4805 state_mem.go:35] "Initializing new in-memory state store" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.244323 4805 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.269740 4805 manager.go:334] "Starting Device Plugin manager" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.269818 4805 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.269834 4805 server.go:79] "Starting device plugin registration server" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.270276 4805 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.270301 4805 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.270520 4805 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.270674 4805 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.270690 4805 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.278481 4805 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.303844 4805 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.304018 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.305054 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.305086 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.305094 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.305195 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.305407 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.305474 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.305975 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.306024 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.306036 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.306224 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.306349 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.306393 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.306898 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.306947 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.306960 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.307253 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.307280 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.307324 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.307333 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.307381 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.307399 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.307553 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.307686 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.307727 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.308271 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.308306 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.308317 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.308478 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.308561 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.308593 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.308667 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.308690 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.308702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.310035 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.310061 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.310074 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.310080 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.310100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.310109 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.310277 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.310307 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.311217 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.311254 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.311274 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.344928 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="400ms" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.365632 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.365734 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.365770 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.365810 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.365847 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.365886 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.365954 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.366003 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.366034 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.366077 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.366110 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.366132 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.366207 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.366245 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.366277 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.370719 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.372018 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.372059 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.372069 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.372093 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.372605 4805 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.196:6443: connect: connection refused" node="crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.467781 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.467848 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.467872 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.467911 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.467932 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.467956 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.467977 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.467996 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468019 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468038 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468058 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468077 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468097 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468117 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468137 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468127 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468285 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468419 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468482 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468513 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468485 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468522 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468545 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468568 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468596 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468606 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468652 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468675 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468682 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.468729 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.573481 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.574521 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.574554 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.574563 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.574581 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.574872 4805 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.196:6443: connect: connection refused" node="crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.637722 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.643664 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.659803 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-85030785b35bc5d007cd60cbc4a981699861c6ebed7fe68144ed745994942303 WatchSource:0}: Error finding container 85030785b35bc5d007cd60cbc4a981699861c6ebed7fe68144ed745994942303: Status 404 returned error can't find the container with id 85030785b35bc5d007cd60cbc4a981699861c6ebed7fe68144ed745994942303 Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.663558 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.665510 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-6b392dc5e1d6b5b6fc009740e7391c52eea1823afc431e9dfb7062d8992bfdec WatchSource:0}: Error finding container 6b392dc5e1d6b5b6fc009740e7391c52eea1823afc431e9dfb7062d8992bfdec: Status 404 returned error can't find the container with id 6b392dc5e1d6b5b6fc009740e7391c52eea1823afc431e9dfb7062d8992bfdec Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.675763 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-c24723337e2925ea47cd9132820dd4a5d9b791a389c306f36c433bc130607c27 WatchSource:0}: Error finding container c24723337e2925ea47cd9132820dd4a5d9b791a389c306f36c433bc130607c27: Status 404 returned error can't find the container with id c24723337e2925ea47cd9132820dd4a5d9b791a389c306f36c433bc130607c27 Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.680824 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.689564 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:26:13 crc kubenswrapper[4805]: W1128 15:26:13.704250 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-401730e70dadc8537bc6a0c358852f7f6beb54414bc0b840c668c67332910ae9 WatchSource:0}: Error finding container 401730e70dadc8537bc6a0c358852f7f6beb54414bc0b840c668c67332910ae9: Status 404 returned error can't find the container with id 401730e70dadc8537bc6a0c358852f7f6beb54414bc0b840c668c67332910ae9 Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.746149 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="800ms" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.975567 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.976963 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.976987 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.976994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:13 crc kubenswrapper[4805]: I1128 15:26:13.977016 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 15:26:13 crc kubenswrapper[4805]: E1128 15:26:13.977453 4805 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.196:6443: connect: connection refused" node="crc" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.141501 4805 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:14 crc kubenswrapper[4805]: W1128 15:26:14.158148 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:14 crc kubenswrapper[4805]: E1128 15:26:14.158221 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.196:6443: connect: connection refused" logger="UnhandledError" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.208699 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29" exitCode=0 Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.208784 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29"} Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.208883 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c24723337e2925ea47cd9132820dd4a5d9b791a389c306f36c433bc130607c27"} Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.208981 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.209904 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.209934 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.209945 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.210524 4805 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944" exitCode=0 Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.210585 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944"} Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.210604 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6b392dc5e1d6b5b6fc009740e7391c52eea1823afc431e9dfb7062d8992bfdec"} Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.210682 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.211503 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.211537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.211575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.211590 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.212511 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.212558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.212571 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.212964 4805 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6c10fa94785110aee1f90c069bbc2aa7ac5ab39aa6382988c291972ddea33deb" exitCode=0 Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.213122 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6c10fa94785110aee1f90c069bbc2aa7ac5ab39aa6382988c291972ddea33deb"} Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.213158 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"85030785b35bc5d007cd60cbc4a981699861c6ebed7fe68144ed745994942303"} Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.213287 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.215651 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.215691 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.215707 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.220077 4805 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e" exitCode=0 Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.220224 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e"} Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.220292 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"401730e70dadc8537bc6a0c358852f7f6beb54414bc0b840c668c67332910ae9"} Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.220420 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.222756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.222800 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.222815 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.224406 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18"} Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.224472 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5e321ab236d7b3fdd0d9631ed3930302b98d8bd55fa0152a211fd03bff210e8f"} Nov 28 15:26:14 crc kubenswrapper[4805]: W1128 15:26:14.352623 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:14 crc kubenswrapper[4805]: E1128 15:26:14.352693 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.196:6443: connect: connection refused" logger="UnhandledError" Nov 28 15:26:14 crc kubenswrapper[4805]: W1128 15:26:14.380106 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:14 crc kubenswrapper[4805]: E1128 15:26:14.380184 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.196:6443: connect: connection refused" logger="UnhandledError" Nov 28 15:26:14 crc kubenswrapper[4805]: W1128 15:26:14.500085 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.196:6443: connect: connection refused Nov 28 15:26:14 crc kubenswrapper[4805]: E1128 15:26:14.501376 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.196:6443: connect: connection refused" logger="UnhandledError" Nov 28 15:26:14 crc kubenswrapper[4805]: E1128 15:26:14.546839 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="1.6s" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.777894 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.779367 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.779396 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.779411 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:14 crc kubenswrapper[4805]: I1128 15:26:14.779432 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.227988 4805 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa" exitCode=0 Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.228054 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.228174 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.229249 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.229274 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.229284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.231176 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"649e8e6f083503baab3d95c91d1f7e86d02a1e6521b91f703fef65d6d7820803"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.231282 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.232064 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.232087 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.232096 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.233938 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.233979 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.234002 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.234083 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.234756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.234774 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.234782 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.236265 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.236298 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.236310 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.236387 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.236937 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.236956 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.236964 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.239496 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.239512 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.239521 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.239530 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.239537 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1"} Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.239591 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.240169 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.240187 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:15 crc kubenswrapper[4805]: I1128 15:26:15.240196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.196286 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.243781 4805 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5" exitCode=0 Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.243869 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5"} Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.243919 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.243960 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.244032 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.244068 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.245280 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.245312 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.245321 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.246124 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.246145 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.246155 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.246227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.246265 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.246278 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:16 crc kubenswrapper[4805]: I1128 15:26:16.420942 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.250845 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f"} Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.250925 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073"} Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.250940 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac"} Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.250949 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12"} Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.250966 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.250983 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.251017 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.252047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.252076 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.252087 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.252898 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.252936 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.252946 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.871412 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.929123 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:17 crc kubenswrapper[4805]: I1128 15:26:17.936090 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.042565 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.042828 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.045236 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.045310 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.045334 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.256876 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883"} Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.256896 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.256954 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.257994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.258031 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.258042 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.258107 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.258136 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:18 crc kubenswrapper[4805]: I1128 15:26:18.258148 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.260266 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.260783 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.260327 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.262452 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.262511 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.262536 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.263350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.263434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.263457 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.382908 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.383167 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.383231 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.384887 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.384952 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:19 crc kubenswrapper[4805]: I1128 15:26:19.384971 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.377344 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.377554 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.378955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.379129 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.379289 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.769354 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.769705 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.771456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.771526 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.771545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.871770 4805 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 15:26:20 crc kubenswrapper[4805]: I1128 15:26:20.871847 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.074398 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.074572 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.076100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.076148 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.076163 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.522681 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.523075 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.525002 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.525029 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:22 crc kubenswrapper[4805]: I1128 15:26:22.525037 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:23 crc kubenswrapper[4805]: E1128 15:26:23.278674 4805 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 15:26:24 crc kubenswrapper[4805]: E1128 15:26:24.780267 4805 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Nov 28 15:26:25 crc kubenswrapper[4805]: I1128 15:26:25.141752 4805 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 28 15:26:25 crc kubenswrapper[4805]: I1128 15:26:25.634992 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 15:26:25 crc kubenswrapper[4805]: I1128 15:26:25.635054 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 15:26:25 crc kubenswrapper[4805]: I1128 15:26:25.658050 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 15:26:25 crc kubenswrapper[4805]: I1128 15:26:25.658128 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 15:26:26 crc kubenswrapper[4805]: I1128 15:26:26.380776 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:26 crc kubenswrapper[4805]: I1128 15:26:26.381964 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:26 crc kubenswrapper[4805]: I1128 15:26:26.382007 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:26 crc kubenswrapper[4805]: I1128 15:26:26.382018 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:26 crc kubenswrapper[4805]: I1128 15:26:26.382041 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 15:26:29 crc kubenswrapper[4805]: I1128 15:26:29.389945 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:29 crc kubenswrapper[4805]: I1128 15:26:29.390717 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:29 crc kubenswrapper[4805]: I1128 15:26:29.392050 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:29 crc kubenswrapper[4805]: I1128 15:26:29.392100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:29 crc kubenswrapper[4805]: I1128 15:26:29.392112 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:29 crc kubenswrapper[4805]: I1128 15:26:29.398412 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.292075 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.293004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.293045 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.293080 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.381162 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.381315 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.382602 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.382631 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.382639 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:30 crc kubenswrapper[4805]: E1128 15:26:30.647778 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.650959 4805 trace.go:236] Trace[1021523090]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 15:26:16.696) (total time: 13953ms): Nov 28 15:26:30 crc kubenswrapper[4805]: Trace[1021523090]: ---"Objects listed" error: 13953ms (15:26:30.650) Nov 28 15:26:30 crc kubenswrapper[4805]: Trace[1021523090]: [13.953918519s] [13.953918519s] END Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.651023 4805 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.651679 4805 trace.go:236] Trace[585272462]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 15:26:16.917) (total time: 13733ms): Nov 28 15:26:30 crc kubenswrapper[4805]: Trace[585272462]: ---"Objects listed" error: 13733ms (15:26:30.651) Nov 28 15:26:30 crc kubenswrapper[4805]: Trace[585272462]: [13.733856772s] [13.733856772s] END Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.651731 4805 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.651940 4805 trace.go:236] Trace[990634094]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 15:26:16.810) (total time: 13841ms): Nov 28 15:26:30 crc kubenswrapper[4805]: Trace[990634094]: ---"Objects listed" error: 13841ms (15:26:30.651) Nov 28 15:26:30 crc kubenswrapper[4805]: Trace[990634094]: [13.841755658s] [13.841755658s] END Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.651961 4805 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.652308 4805 trace.go:236] Trace[1441952558]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 15:26:17.095) (total time: 13556ms): Nov 28 15:26:30 crc kubenswrapper[4805]: Trace[1441952558]: ---"Objects listed" error: 13556ms (15:26:30.652) Nov 28 15:26:30 crc kubenswrapper[4805]: Trace[1441952558]: [13.55654255s] [13.55654255s] END Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.652338 4805 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.652648 4805 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 28 15:26:30 crc kubenswrapper[4805]: I1128 15:26:30.816790 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.149451 4805 apiserver.go:52] "Watching apiserver" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.151866 4805 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.152301 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.152814 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.152894 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.152936 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.152939 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.153051 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.153273 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.153298 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.153330 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.153626 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.158486 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.160188 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.160389 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.160522 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.160615 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.160799 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.161071 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.161086 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.162112 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.185083 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.199758 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.202938 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43318->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.202995 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43322->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.202990 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43318->192.168.126.11:17697: read: connection reset by peer" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.203056 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43322->192.168.126.11:17697: read: connection reset by peer" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.203485 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.203587 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.204296 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.204332 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.215034 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.227430 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.249649 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.256321 4805 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258679 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258745 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258768 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258793 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258814 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258835 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258857 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258882 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258904 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258924 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258947 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258968 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.258986 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259010 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259029 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259046 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259065 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259086 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259105 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259126 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259144 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259163 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259182 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259199 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259217 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259237 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259255 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259276 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259295 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259316 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259334 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259373 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259394 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259415 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259440 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259460 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259481 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259500 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259520 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259538 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259557 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259576 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259595 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259615 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259634 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259674 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259693 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259713 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259733 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259752 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259771 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259792 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259813 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259834 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259854 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259872 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259890 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259909 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259929 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259948 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.259967 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260017 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260037 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260059 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260079 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260100 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260120 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260140 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260162 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260181 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260203 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260223 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260243 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260263 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260283 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260306 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260329 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260351 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260391 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260412 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260432 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260452 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260471 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260494 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260513 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260563 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260585 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260605 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260627 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260649 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260670 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260692 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260711 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260734 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260754 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260739 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260776 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260799 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260820 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260843 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260868 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260887 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260907 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260927 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260948 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260969 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.260990 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261012 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261035 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261056 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261102 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261124 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261147 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261170 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261191 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261214 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261238 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261260 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261283 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261305 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261328 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261351 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261386 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261394 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261451 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261479 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261524 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261551 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261578 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261604 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261622 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261646 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261672 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261695 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261689 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261715 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261780 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261805 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261827 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261848 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261868 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261883 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261914 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261933 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261953 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.261975 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262017 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262034 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262053 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262068 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262085 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262101 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262117 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262135 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262157 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262177 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262195 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262203 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262213 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262263 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262284 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262304 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262321 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262343 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262374 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262378 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262418 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262438 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262456 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262474 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262493 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262504 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262512 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262643 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262648 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262672 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262700 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262733 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262756 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262783 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262789 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262809 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262837 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262860 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262860 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262885 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262909 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262935 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262961 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262984 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263010 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263032 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263056 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263079 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263187 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263218 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263240 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263264 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263286 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263307 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263329 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263371 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263399 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263422 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263470 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263498 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263524 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263549 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263571 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263597 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263618 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263645 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263704 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263729 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263757 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263781 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263807 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263827 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263880 4805 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263896 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263910 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263922 4805 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263933 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263946 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263957 4805 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263969 4805 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.270060 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.262961 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263041 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263085 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.278896 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263171 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263233 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263512 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.263845 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.264163 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.264312 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.264781 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.265009 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.265200 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.265643 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.265861 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.265884 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.266093 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.266103 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.266341 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.266592 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.266705 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.267148 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.267328 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.267508 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.267961 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.268254 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.268752 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.269640 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.269895 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.269900 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.270139 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.270283 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.270520 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.270673 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.270745 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.270933 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271076 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271083 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271108 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271141 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271179 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271370 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271393 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271695 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271729 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271844 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.271895 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272050 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272197 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272206 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272395 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272549 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272621 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272760 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272800 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272712 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.272983 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.273012 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.273102 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.273226 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.273274 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.273438 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.273469 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.273566 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.273604 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.274420 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.274486 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.274589 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:26:31.774567723 +0000 UTC m=+18.824359034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.274591 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.274947 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.274954 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.275639 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.275682 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.275697 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.275793 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.275830 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.275114 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.275994 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.276529 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.276574 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.276593 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.276803 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.277156 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.277173 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.277179 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.277235 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.277626 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.277705 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.277710 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.277889 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.277932 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.278086 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.278465 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.278635 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.278642 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.278695 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.278984 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.279027 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.279039 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.278478 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.279331 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.279653 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.279751 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.279867 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.279978 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.280307 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.280713 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.281605 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.281924 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.283273 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.283373 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.285665 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.285917 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.286242 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.286468 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.286511 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.286694 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.286842 4805 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.286923 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.287101 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.287294 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.288711 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.288734 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.288994 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.289099 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.289152 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.289406 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.289513 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.289796 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.289982 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.290241 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.290280 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.290701 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.290703 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.290948 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.291186 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.291255 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.291424 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.292049 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.292049 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.292142 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.292372 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.292417 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.293590 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.293741 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.293839 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.294073 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.294311 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.294501 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.294559 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.295054 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.295135 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.295283 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:31.795258416 +0000 UTC m=+18.845049727 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.295527 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:31.795516955 +0000 UTC m=+18.845308266 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.296592 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.296778 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.296867 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.296486 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.295948 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.296042 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.296122 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.296146 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.296468 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.297119 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.300454 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.300802 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.300936 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.301105 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.301507 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.301633 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.302141 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.302441 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.302526 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.302897 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.302985 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.303060 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.303568 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.304368 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.303801 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.304500 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.303903 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.304441 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.304721 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:31.804692652 +0000 UTC m=+18.854483963 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.305344 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.305449 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.305817 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.306712 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.311781 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.311959 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.312494 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.316138 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.316432 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.318010 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.318080 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.318101 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.318174 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:31.818149795 +0000 UTC m=+18.867941106 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.318382 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.318383 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.323002 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.324511 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.324699 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.324707 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.324766 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68" exitCode=255 Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.324850 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.329712 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.331580 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.332769 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.335189 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.335265 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.338125 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.339180 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.342163 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.343018 4805 scope.go:117] "RemoveContainer" containerID="b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.344178 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.349800 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.354716 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.365846 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.366453 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.366567 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.374752 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.374971 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.374996 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375008 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375020 4805 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375030 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375039 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375050 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375064 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375075 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375085 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375094 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375105 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375115 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375124 4805 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375134 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375143 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375153 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375164 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375175 4805 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375184 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375194 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375204 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375214 4805 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375226 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375236 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375247 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375259 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375270 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375281 4805 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375291 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375300 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375310 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375320 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375329 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375339 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375349 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375383 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375393 4805 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375403 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375413 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375423 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375433 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375444 4805 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375453 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375462 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375473 4805 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375485 4805 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375496 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375479 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375507 4805 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375599 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375630 4805 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375643 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375656 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375669 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375679 4805 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375690 4805 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375702 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375713 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375725 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375735 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375746 4805 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375756 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375767 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375777 4805 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375787 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375798 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375811 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375824 4805 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375836 4805 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375849 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375863 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375876 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375890 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375905 4805 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375919 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375933 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375945 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.375955 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376164 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376175 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376187 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376197 4805 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376210 4805 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376221 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376233 4805 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376244 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376256 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376267 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376278 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376291 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376303 4805 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376315 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376326 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376337 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376385 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376396 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376406 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376418 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376429 4805 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376439 4805 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376449 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376459 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376470 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376480 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376490 4805 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376500 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376511 4805 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376521 4805 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376533 4805 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376543 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376554 4805 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376564 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376577 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376588 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376601 4805 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376612 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376625 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376636 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376647 4805 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376657 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376667 4805 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376677 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376688 4805 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376698 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376713 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376725 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376739 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376752 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376727 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376765 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376897 4805 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376908 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376918 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376929 4805 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376940 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376950 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376960 4805 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376970 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376983 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.376995 4805 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377005 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377017 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377027 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377041 4805 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377052 4805 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377062 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377074 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377085 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377095 4805 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377107 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377117 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377128 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377137 4805 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377148 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377159 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377169 4805 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377182 4805 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377223 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377234 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377247 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377259 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377279 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377289 4805 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377300 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377332 4805 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377344 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377369 4805 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377380 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377391 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377402 4805 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377417 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377429 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377439 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377449 4805 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377460 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377471 4805 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377483 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377493 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377503 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377514 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377524 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377536 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377546 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377557 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377570 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377579 4805 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.377589 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.389558 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.391291 4805 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.391419 4805 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.392894 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.392918 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.392926 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.392942 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.392965 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.400478 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.408314 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.411719 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.427061 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.429477 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.429506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.429516 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.429533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.429545 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.441610 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.444002 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.446864 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.446914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.446925 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.446946 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.446959 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.455014 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.457918 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.465171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.465240 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.465256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.465279 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.465292 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.466513 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.474963 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.481610 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.493497 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.494604 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.494638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.494646 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.494663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.494673 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.499710 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.502801 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.508654 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.508867 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.518801 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.518849 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.518860 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.518880 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.518898 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.522597 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: W1128 15:26:31.533985 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-9cd0634eceb4923e9c67b60221f617fde56871a989f6dbcd32883bfca0e45ffb WatchSource:0}: Error finding container 9cd0634eceb4923e9c67b60221f617fde56871a989f6dbcd32883bfca0e45ffb: Status 404 returned error can't find the container with id 9cd0634eceb4923e9c67b60221f617fde56871a989f6dbcd32883bfca0e45ffb Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.545621 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.559131 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.623546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.623593 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.623606 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.623626 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.623638 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.725340 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.725419 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.725431 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.725448 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.725458 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.783998 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.784194 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:26:32.784166316 +0000 UTC m=+19.833957627 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.806770 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-9tjqf"] Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.807081 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9tjqf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.808808 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.810410 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.821512 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.825066 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.828463 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.828507 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.828518 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.828537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.828547 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.844089 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.863617 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.881849 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.885512 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f70dd0f8-a9a1-4694-827a-6c27deed4080-hosts-file\") pod \"node-resolver-9tjqf\" (UID: \"f70dd0f8-a9a1-4694-827a-6c27deed4080\") " pod="openshift-dns/node-resolver-9tjqf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.885568 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.885595 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.885623 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mkgd\" (UniqueName: \"kubernetes.io/projected/f70dd0f8-a9a1-4694-827a-6c27deed4080-kube-api-access-9mkgd\") pod \"node-resolver-9tjqf\" (UID: \"f70dd0f8-a9a1-4694-827a-6c27deed4080\") " pod="openshift-dns/node-resolver-9tjqf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.885650 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.885677 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.885807 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.885826 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.885840 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.885887 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:32.885872252 +0000 UTC m=+19.935663573 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.886137 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.886176 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:32.88616783 +0000 UTC m=+19.935959141 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.886233 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.886245 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.886255 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.886295 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:32.886281043 +0000 UTC m=+19.936072354 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.886386 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: E1128 15:26:31.886422 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:32.886414288 +0000 UTC m=+19.936205599 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.910024 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.930418 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.930466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.930478 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.930500 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.930514 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:31Z","lastTransitionTime":"2025-11-28T15:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.932454 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.967445 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.984285 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.986556 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f70dd0f8-a9a1-4694-827a-6c27deed4080-hosts-file\") pod \"node-resolver-9tjqf\" (UID: \"f70dd0f8-a9a1-4694-827a-6c27deed4080\") " pod="openshift-dns/node-resolver-9tjqf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.986647 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mkgd\" (UniqueName: \"kubernetes.io/projected/f70dd0f8-a9a1-4694-827a-6c27deed4080-kube-api-access-9mkgd\") pod \"node-resolver-9tjqf\" (UID: \"f70dd0f8-a9a1-4694-827a-6c27deed4080\") " pod="openshift-dns/node-resolver-9tjqf" Nov 28 15:26:31 crc kubenswrapper[4805]: I1128 15:26:31.986705 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f70dd0f8-a9a1-4694-827a-6c27deed4080-hosts-file\") pod \"node-resolver-9tjqf\" (UID: \"f70dd0f8-a9a1-4694-827a-6c27deed4080\") " pod="openshift-dns/node-resolver-9tjqf" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.015109 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.032927 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.032967 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.032976 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.032991 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.033000 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.109772 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mkgd\" (UniqueName: \"kubernetes.io/projected/f70dd0f8-a9a1-4694-827a-6c27deed4080-kube-api-access-9mkgd\") pod \"node-resolver-9tjqf\" (UID: \"f70dd0f8-a9a1-4694-827a-6c27deed4080\") " pod="openshift-dns/node-resolver-9tjqf" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.118143 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9tjqf" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.135491 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.135524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.135533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.135547 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.135556 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.238184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.238498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.238564 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.238649 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.238753 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.284681 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-9pzmp"] Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.285329 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.289718 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.290088 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.290456 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.290667 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.300453 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.318185 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.329246 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9tjqf" event={"ID":"f70dd0f8-a9a1-4694-827a-6c27deed4080","Type":"ContainerStarted","Data":"34354480e8878b0f5d716f92b11019925fbb92b475aa8ef6ffc2614678e837be"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.330365 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9cd0634eceb4923e9c67b60221f617fde56871a989f6dbcd32883bfca0e45ffb"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.332700 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.332754 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.332769 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2369ca9264517d60840834082a6005b2a3a7d2e50c52dfb9d314a96b48970342"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.335444 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.336737 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.337265 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.338991 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.339016 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1820971e6492260889c1fec5effdd8d3424c564ebfdc7d90428b82a0327b0607"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.340473 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.340500 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.340512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.340525 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.340535 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.340759 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.358124 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.371047 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.383318 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.389201 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/73da70d8-a7dc-4fca-9e65-9c0d0d815966-rootfs\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.389262 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftdf5\" (UniqueName: \"kubernetes.io/projected/73da70d8-a7dc-4fca-9e65-9c0d0d815966-kube-api-access-ftdf5\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.389313 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/73da70d8-a7dc-4fca-9e65-9c0d0d815966-proxy-tls\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.389339 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/73da70d8-a7dc-4fca-9e65-9c0d0d815966-mcd-auth-proxy-config\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.417725 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.443116 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.443174 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.443185 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.443218 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.443233 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.464740 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.489490 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.489951 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/73da70d8-a7dc-4fca-9e65-9c0d0d815966-rootfs\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.490013 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftdf5\" (UniqueName: \"kubernetes.io/projected/73da70d8-a7dc-4fca-9e65-9c0d0d815966-kube-api-access-ftdf5\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.490066 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/73da70d8-a7dc-4fca-9e65-9c0d0d815966-proxy-tls\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.490092 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/73da70d8-a7dc-4fca-9e65-9c0d0d815966-mcd-auth-proxy-config\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.490065 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/73da70d8-a7dc-4fca-9e65-9c0d0d815966-rootfs\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.490901 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/73da70d8-a7dc-4fca-9e65-9c0d0d815966-mcd-auth-proxy-config\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.494665 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/73da70d8-a7dc-4fca-9e65-9c0d0d815966-proxy-tls\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.516928 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.521616 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftdf5\" (UniqueName: \"kubernetes.io/projected/73da70d8-a7dc-4fca-9e65-9c0d0d815966-kube-api-access-ftdf5\") pod \"machine-config-daemon-9pzmp\" (UID: \"73da70d8-a7dc-4fca-9e65-9c0d0d815966\") " pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.543663 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.545787 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.545834 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.545846 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.545865 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.545880 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.567151 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.574792 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.589478 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.602737 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.604791 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.611245 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.628225 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.643791 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.646978 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.649320 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.649377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.649388 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.649406 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.649417 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.668952 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.679404 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-fv2dw"] Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.680020 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.684909 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.684924 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.684943 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.684963 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.685055 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6v4kb"] Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.685303 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.685858 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.688573 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.689054 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.689064 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.689287 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.693009 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.693307 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.693352 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.695569 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-wdzqc"] Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.696474 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.698183 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.702697 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.703020 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.721151 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.742752 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.755472 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.755523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.755544 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.755564 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.755577 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.761992 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.782566 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793158 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793263 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpjph\" (UniqueName: \"kubernetes.io/projected/412627f3-6ef3-401d-b7ae-a839d70a46b3-kube-api-access-fpjph\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793288 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45ztm\" (UniqueName: \"kubernetes.io/projected/9d0031c5-0433-419e-9363-66eb48341a68-kube-api-access-45ztm\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793307 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-cnibin\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793322 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-multus-socket-dir-parent\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793338 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-var-lib-cni-multus\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793374 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-system-cni-dir\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.793458 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:26:34.793374299 +0000 UTC m=+21.843165610 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793544 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-bin\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793623 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-systemd\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793716 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-netd\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793746 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-var-lib-cni-bin\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793768 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-hostroot\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793791 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-run-multus-certs\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793863 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-os-release\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793889 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793923 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793947 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-system-cni-dir\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793972 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-kubelet\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.793993 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-systemd-units\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794015 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-slash\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794039 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-var-lib-openvswitch\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794065 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-etc-kubernetes\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794085 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-multus-conf-dir\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794105 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-run-k8s-cni-cncf-io\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794120 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-script-lib\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794135 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-cnibin\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794151 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e746e8b8-a375-4650-98fc-f6e3ba35a50d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794169 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e746e8b8-a375-4650-98fc-f6e3ba35a50d-cni-binary-copy\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794188 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mf4k\" (UniqueName: \"kubernetes.io/projected/e746e8b8-a375-4650-98fc-f6e3ba35a50d-kube-api-access-9mf4k\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794221 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-var-lib-kubelet\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794241 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-node-log\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794262 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-run-netns\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794282 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9d0031c5-0433-419e-9363-66eb48341a68-multus-daemon-config\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794302 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-etc-openvswitch\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794608 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovn-node-metrics-cert\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794641 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-multus-cni-dir\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794664 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9d0031c5-0433-419e-9363-66eb48341a68-cni-binary-copy\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794720 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-ovn\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794740 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-log-socket\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794787 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-config\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794804 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-os-release\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794875 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-openvswitch\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794898 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-netns\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794945 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-env-overrides\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.794963 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.798900 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.822819 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.841394 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.858634 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.858683 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.858695 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.858713 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.858725 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.860936 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.873922 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.892153 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.896664 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45ztm\" (UniqueName: \"kubernetes.io/projected/9d0031c5-0433-419e-9363-66eb48341a68-kube-api-access-45ztm\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.896745 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpjph\" (UniqueName: \"kubernetes.io/projected/412627f3-6ef3-401d-b7ae-a839d70a46b3-kube-api-access-fpjph\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.896848 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-system-cni-dir\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.896880 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-cnibin\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.896911 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-multus-socket-dir-parent\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.896939 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-var-lib-cni-multus\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.896973 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-bin\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897003 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-systemd\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897032 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-netd\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897062 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-os-release\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897091 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-var-lib-cni-bin\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897109 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-var-lib-cni-multus\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897131 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-bin\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897119 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-hostroot\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897147 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-cnibin\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897199 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-os-release\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897189 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-multus-socket-dir-parent\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897235 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-hostroot\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897240 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-system-cni-dir\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897217 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-run-multus-certs\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897222 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-systemd\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897205 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-var-lib-cni-bin\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897297 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-run-multus-certs\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897189 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-netd\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897394 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897426 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897450 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897470 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-system-cni-dir\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897491 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-var-lib-openvswitch\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897516 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-kubelet\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897514 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897490 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897539 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-systemd-units\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.897554 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.897576 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.897590 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897617 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-slash\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897520 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-system-cni-dir\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897588 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-slash\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897640 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-kubelet\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897651 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-var-lib-openvswitch\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897562 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-systemd-units\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.897662 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:34.89764326 +0000 UTC m=+21.947434761 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897754 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-etc-kubernetes\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897785 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-multus-conf-dir\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897810 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-run-k8s-cni-cncf-io\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897807 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-etc-kubernetes\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897866 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-run-k8s-cni-cncf-io\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897868 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-multus-conf-dir\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897830 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-script-lib\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897923 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-cnibin\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897948 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e746e8b8-a375-4650-98fc-f6e3ba35a50d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897956 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-cnibin\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897973 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-node-log\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.897999 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e746e8b8-a375-4650-98fc-f6e3ba35a50d-cni-binary-copy\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898020 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mf4k\" (UniqueName: \"kubernetes.io/projected/e746e8b8-a375-4650-98fc-f6e3ba35a50d-kube-api-access-9mf4k\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898045 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-node-log\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898048 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898086 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-var-lib-kubelet\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.898098 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898108 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-run-netns\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898130 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9d0031c5-0433-419e-9363-66eb48341a68-multus-daemon-config\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.898152 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:34.898129755 +0000 UTC m=+21.947921276 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898151 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-var-lib-kubelet\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898169 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-etc-openvswitch\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898193 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-log-socket\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898213 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovn-node-metrics-cert\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898222 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-etc-openvswitch\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898234 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-multus-cni-dir\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898255 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9d0031c5-0433-419e-9363-66eb48341a68-cni-binary-copy\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898277 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-ovn\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898302 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-openvswitch\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898343 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-config\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898344 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-ovn\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898291 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-log-socket\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898416 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-os-release\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898433 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-host-run-netns\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898525 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898535 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-openvswitch\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898559 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9d0031c5-0433-419e-9363-66eb48341a68-multus-cni-dir\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898640 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.898686 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.898708 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.898720 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.898830 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:34.898799684 +0000 UTC m=+21.948590995 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.898878 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898907 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898939 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-netns\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: E1128 15:26:32.898969 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:34.898939158 +0000 UTC m=+21.948730699 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898982 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-netns\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.898998 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-env-overrides\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.899124 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9d0031c5-0433-419e-9363-66eb48341a68-cni-binary-copy\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.899173 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-script-lib\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.899175 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-config\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.899224 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-os-release\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.899382 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e746e8b8-a375-4650-98fc-f6e3ba35a50d-cni-binary-copy\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.899396 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9d0031c5-0433-419e-9363-66eb48341a68-multus-daemon-config\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.899398 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e746e8b8-a375-4650-98fc-f6e3ba35a50d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.899732 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e746e8b8-a375-4650-98fc-f6e3ba35a50d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.899889 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-env-overrides\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.905516 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovn-node-metrics-cert\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.914230 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.919691 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpjph\" (UniqueName: \"kubernetes.io/projected/412627f3-6ef3-401d-b7ae-a839d70a46b3-kube-api-access-fpjph\") pod \"ovnkube-node-6v4kb\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.919994 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mf4k\" (UniqueName: \"kubernetes.io/projected/e746e8b8-a375-4650-98fc-f6e3ba35a50d-kube-api-access-9mf4k\") pod \"multus-additional-cni-plugins-wdzqc\" (UID: \"e746e8b8-a375-4650-98fc-f6e3ba35a50d\") " pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.922160 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45ztm\" (UniqueName: \"kubernetes.io/projected/9d0031c5-0433-419e-9363-66eb48341a68-kube-api-access-45ztm\") pod \"multus-fv2dw\" (UID: \"9d0031c5-0433-419e-9363-66eb48341a68\") " pod="openshift-multus/multus-fv2dw" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.934261 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.960922 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.960978 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.960993 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.961022 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.961038 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:32Z","lastTransitionTime":"2025-11-28T15:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.970872 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:32 crc kubenswrapper[4805]: I1128 15:26:32.997719 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fv2dw" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.013529 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:33 crc kubenswrapper[4805]: W1128 15:26:33.013680 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d0031c5_0433_419e_9363_66eb48341a68.slice/crio-11b6c9ad5c94f7165c4dc357973b9537fc92ceb3cc04f620238dd5b19af15eef WatchSource:0}: Error finding container 11b6c9ad5c94f7165c4dc357973b9537fc92ceb3cc04f620238dd5b19af15eef: Status 404 returned error can't find the container with id 11b6c9ad5c94f7165c4dc357973b9537fc92ceb3cc04f620238dd5b19af15eef Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.020760 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.025568 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: W1128 15:26:33.041425 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod412627f3_6ef3_401d_b7ae_a839d70a46b3.slice/crio-52d2a7adc518626b5899bd59433b5ecaf2b00cfb66b7e4c01657d37b37955b48 WatchSource:0}: Error finding container 52d2a7adc518626b5899bd59433b5ecaf2b00cfb66b7e4c01657d37b37955b48: Status 404 returned error can't find the container with id 52d2a7adc518626b5899bd59433b5ecaf2b00cfb66b7e4c01657d37b37955b48 Nov 28 15:26:33 crc kubenswrapper[4805]: W1128 15:26:33.051932 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode746e8b8_a375_4650_98fc_f6e3ba35a50d.slice/crio-b6be4633d4e1aa09940dbb77ada15ebff2938d8d0755f3f16992e81ddea8a8e3 WatchSource:0}: Error finding container b6be4633d4e1aa09940dbb77ada15ebff2938d8d0755f3f16992e81ddea8a8e3: Status 404 returned error can't find the container with id b6be4633d4e1aa09940dbb77ada15ebff2938d8d0755f3f16992e81ddea8a8e3 Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.064426 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.064460 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.064469 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.064485 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.064495 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.068671 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.098482 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.119027 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.167792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.167830 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.167839 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.167855 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.167865 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.203206 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.203235 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.203330 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:33 crc kubenswrapper[4805]: E1128 15:26:33.203491 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:33 crc kubenswrapper[4805]: E1128 15:26:33.203610 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:33 crc kubenswrapper[4805]: E1128 15:26:33.203718 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.207697 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.208459 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.215833 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.217071 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.218272 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.218885 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.219585 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.220716 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.226150 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.226843 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.227947 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.228275 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.229051 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.230146 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.230690 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.231236 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.232519 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.233135 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.234093 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.234721 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.235440 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.236512 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.237074 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.237550 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.240163 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.241332 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.242036 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.243218 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.243828 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.244886 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.245453 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.247518 4805 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.247772 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.251281 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.252327 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.252916 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.254802 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.256042 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.257598 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.258262 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.259522 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.260395 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.261514 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.262192 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.263312 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.264324 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.264817 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.265345 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.266293 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.267048 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.268092 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.268660 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.269606 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.270329 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.270554 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.270590 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.270602 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.270620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.270632 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.270990 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.271971 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.273418 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.288234 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.311949 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.330936 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.362893 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.365008 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9tjqf" event={"ID":"f70dd0f8-a9a1-4694-827a-6c27deed4080","Type":"ContainerStarted","Data":"b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.371517 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff" exitCode=0 Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.371599 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.371649 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"52d2a7adc518626b5899bd59433b5ecaf2b00cfb66b7e4c01657d37b37955b48"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.373648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.374069 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.374084 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.374105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.374118 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.374750 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fv2dw" event={"ID":"9d0031c5-0433-419e-9363-66eb48341a68","Type":"ContainerStarted","Data":"64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.374871 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fv2dw" event={"ID":"9d0031c5-0433-419e-9363-66eb48341a68","Type":"ContainerStarted","Data":"11b6c9ad5c94f7165c4dc357973b9537fc92ceb3cc04f620238dd5b19af15eef"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.380997 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" event={"ID":"e746e8b8-a375-4650-98fc-f6e3ba35a50d","Type":"ContainerStarted","Data":"7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.381050 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" event={"ID":"e746e8b8-a375-4650-98fc-f6e3ba35a50d","Type":"ContainerStarted","Data":"b6be4633d4e1aa09940dbb77ada15ebff2938d8d0755f3f16992e81ddea8a8e3"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.383451 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.387984 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.388072 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.388089 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"409a89856a00e3636316f49fe7b34ccd730bce4b573a6c0d1cf66ac15364c78e"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.405303 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.420917 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.437214 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.449850 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.467818 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.476776 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.476821 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.476833 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.476852 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.476865 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.481135 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.502568 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.529035 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.546549 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.567909 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.579437 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.579482 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.579494 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.579512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.579525 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.588349 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.616165 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.642946 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.666322 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.681806 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.681875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.681952 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.681996 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.682028 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.699926 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.715124 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.729416 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.740442 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.772559 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.784623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.784657 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.784665 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.784678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.784687 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.815536 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.887479 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.887511 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.887522 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.887539 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.887559 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.993863 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.993908 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.993919 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.993935 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:33 crc kubenswrapper[4805]: I1128 15:26:33.993948 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:33Z","lastTransitionTime":"2025-11-28T15:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.096582 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.096626 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.096638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.096655 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.096667 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:34Z","lastTransitionTime":"2025-11-28T15:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.199910 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.199955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.199966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.199983 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.199994 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:34Z","lastTransitionTime":"2025-11-28T15:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.302723 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.303189 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.303207 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.303229 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.303243 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:34Z","lastTransitionTime":"2025-11-28T15:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.393715 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.393947 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.394038 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.394096 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.394151 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.394212 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.394795 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.396410 4805 generic.go:334] "Generic (PLEG): container finished" podID="e746e8b8-a375-4650-98fc-f6e3ba35a50d" containerID="7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18" exitCode=0 Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.396451 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" event={"ID":"e746e8b8-a375-4650-98fc-f6e3ba35a50d","Type":"ContainerDied","Data":"7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.405253 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.405297 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.405309 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.405327 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.405343 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:34Z","lastTransitionTime":"2025-11-28T15:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.409043 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.422032 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.437538 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.451600 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.463269 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.475650 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.485679 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.498553 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.508273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.508340 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.508377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.508398 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.508410 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:34Z","lastTransitionTime":"2025-11-28T15:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.515996 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.538687 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.554533 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.567887 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.581122 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.593966 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.606505 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.610475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.610514 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.610524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.610538 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.610548 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:34Z","lastTransitionTime":"2025-11-28T15:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.617565 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.633075 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.645035 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.658146 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.668968 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.681499 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.693714 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.713276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.713444 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.713531 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.713623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.713701 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:34Z","lastTransitionTime":"2025-11-28T15:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.739131 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.779632 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.812573 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.816478 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.816609 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.816712 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.816779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.816838 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:34Z","lastTransitionTime":"2025-11-28T15:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.819878 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.820065 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:26:38.820044318 +0000 UTC m=+25.869835629 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.851159 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.894107 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.918869 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.918912 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.918923 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.918940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.918952 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:34Z","lastTransitionTime":"2025-11-28T15:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.920249 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.920288 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.920307 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.920379 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920407 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920460 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:38.920443157 +0000 UTC m=+25.970234468 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920466 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920467 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920489 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920501 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920526 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920534 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:38.920524579 +0000 UTC m=+25.970315890 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920480 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920551 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920559 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:38.92054904 +0000 UTC m=+25.970340351 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:34 crc kubenswrapper[4805]: E1128 15:26:34.920578 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:38.92056799 +0000 UTC m=+25.970359301 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.933513 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:34Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.945609 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-jmhd2"] Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.946051 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.947608 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 15:26:34 crc kubenswrapper[4805]: I1128 15:26:34.964653 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.005783 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.020798 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh2vb\" (UniqueName: \"kubernetes.io/projected/c3d26ccf-e51d-4b2d-a17f-766fa06e71ff-kube-api-access-nh2vb\") pod \"node-ca-jmhd2\" (UID: \"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\") " pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.020894 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3d26ccf-e51d-4b2d-a17f-766fa06e71ff-serviceca\") pod \"node-ca-jmhd2\" (UID: \"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\") " pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.020916 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c3d26ccf-e51d-4b2d-a17f-766fa06e71ff-host\") pod \"node-ca-jmhd2\" (UID: \"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\") " pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.021669 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.021694 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.021706 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.021722 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.021734 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.024662 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.054406 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.093526 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.121927 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh2vb\" (UniqueName: \"kubernetes.io/projected/c3d26ccf-e51d-4b2d-a17f-766fa06e71ff-kube-api-access-nh2vb\") pod \"node-ca-jmhd2\" (UID: \"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\") " pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.121971 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3d26ccf-e51d-4b2d-a17f-766fa06e71ff-serviceca\") pod \"node-ca-jmhd2\" (UID: \"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\") " pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.121990 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c3d26ccf-e51d-4b2d-a17f-766fa06e71ff-host\") pod \"node-ca-jmhd2\" (UID: \"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\") " pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.122050 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c3d26ccf-e51d-4b2d-a17f-766fa06e71ff-host\") pod \"node-ca-jmhd2\" (UID: \"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\") " pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.123846 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.123883 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.123896 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.123915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.123933 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.130445 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.146079 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3d26ccf-e51d-4b2d-a17f-766fa06e71ff-serviceca\") pod \"node-ca-jmhd2\" (UID: \"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\") " pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.163391 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh2vb\" (UniqueName: \"kubernetes.io/projected/c3d26ccf-e51d-4b2d-a17f-766fa06e71ff-kube-api-access-nh2vb\") pod \"node-ca-jmhd2\" (UID: \"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\") " pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.191962 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.203384 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.203427 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.203472 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:35 crc kubenswrapper[4805]: E1128 15:26:35.203509 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:35 crc kubenswrapper[4805]: E1128 15:26:35.203605 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:35 crc kubenswrapper[4805]: E1128 15:26:35.203691 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.226250 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.226288 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.226297 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.226313 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.226322 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.231900 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.260496 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jmhd2" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.271306 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.312916 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.329464 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.329504 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.329516 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.329536 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.329548 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.355346 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.392827 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.407992 4805 generic.go:334] "Generic (PLEG): container finished" podID="e746e8b8-a375-4650-98fc-f6e3ba35a50d" containerID="6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1" exitCode=0 Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.408093 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" event={"ID":"e746e8b8-a375-4650-98fc-f6e3ba35a50d","Type":"ContainerDied","Data":"6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.409864 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jmhd2" event={"ID":"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff","Type":"ContainerStarted","Data":"5723ecaf645d568f2fecf73b46f6d65a86492b74f9fe8b23058e0b857c5e4c31"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.433716 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.433785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.433804 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.433829 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.433845 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.436253 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.475863 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.514574 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.536467 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.536500 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.536509 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.536521 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.536530 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.553821 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.604896 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.634436 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.639028 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.639091 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.639099 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.639114 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.639123 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.678280 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.715677 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.742241 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.742288 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.742299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.742320 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.742334 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.751397 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.793740 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.835475 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.845247 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.845324 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.845339 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.845374 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.845385 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.871343 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.913812 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.948068 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.948108 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.948119 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.948135 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.948147 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:35Z","lastTransitionTime":"2025-11-28T15:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.951915 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:35 crc kubenswrapper[4805]: I1128 15:26:35.996980 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:35Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.032760 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.050395 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.050447 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.050466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.050486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.050501 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.072267 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.112784 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.152193 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.153562 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.153619 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.153632 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.153655 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.153667 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.194193 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.238201 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.256152 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.256200 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.256213 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.256234 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.256247 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.358312 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.358352 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.358382 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.358400 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.358412 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.416796 4805 generic.go:334] "Generic (PLEG): container finished" podID="e746e8b8-a375-4650-98fc-f6e3ba35a50d" containerID="044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044" exitCode=0 Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.416838 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" event={"ID":"e746e8b8-a375-4650-98fc-f6e3ba35a50d","Type":"ContainerDied","Data":"044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.419061 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jmhd2" event={"ID":"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff","Type":"ContainerStarted","Data":"0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.430914 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.441291 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.455774 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.461730 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.461807 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.461823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.461846 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.461860 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.470109 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.481321 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.494536 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.509877 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.553135 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.564126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.564174 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.564186 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.564203 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.564216 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.597820 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.636920 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.666615 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.666650 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.666660 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.666674 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.666697 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.674019 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.713620 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.757145 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.769659 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.769711 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.769730 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.769784 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.769801 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.796772 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.833505 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.873191 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.873256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.873273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.873298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.873315 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.874757 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.915930 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.954807 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.976255 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.976294 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.976305 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.976323 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.976336 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:36Z","lastTransitionTime":"2025-11-28T15:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:36 crc kubenswrapper[4805]: I1128 15:26:36.992436 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:36Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.035188 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.073247 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.079062 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.079112 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.079132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.079167 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.079181 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:37Z","lastTransitionTime":"2025-11-28T15:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.112730 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.152052 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.186671 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.186719 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.186731 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.186748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.186761 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:37Z","lastTransitionTime":"2025-11-28T15:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.202163 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.203200 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.203274 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.203344 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:37 crc kubenswrapper[4805]: E1128 15:26:37.203345 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:37 crc kubenswrapper[4805]: E1128 15:26:37.203453 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:37 crc kubenswrapper[4805]: E1128 15:26:37.203547 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.234805 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.273013 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.289442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.289480 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.289488 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.289502 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.289511 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:37Z","lastTransitionTime":"2025-11-28T15:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.316800 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.352723 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.391463 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.392042 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.392076 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.392084 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.392099 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.392109 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:37Z","lastTransitionTime":"2025-11-28T15:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.425562 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.428052 4805 generic.go:334] "Generic (PLEG): container finished" podID="e746e8b8-a375-4650-98fc-f6e3ba35a50d" containerID="5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0" exitCode=0 Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.428171 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" event={"ID":"e746e8b8-a375-4650-98fc-f6e3ba35a50d","Type":"ContainerDied","Data":"5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.437573 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.475417 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.505414 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.505456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.505465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.505482 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.505494 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:37Z","lastTransitionTime":"2025-11-28T15:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.522296 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.553903 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.594319 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.608133 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.608181 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.608218 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.608239 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.608251 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:37Z","lastTransitionTime":"2025-11-28T15:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.633490 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.673442 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.711318 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.711382 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.711397 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.711417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.711436 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:37Z","lastTransitionTime":"2025-11-28T15:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.713319 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.759616 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.794266 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.814201 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.814238 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.814247 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.814262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.814273 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:37Z","lastTransitionTime":"2025-11-28T15:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.833448 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.876301 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.913111 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.916791 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.916832 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.916844 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.916887 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.916900 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:37Z","lastTransitionTime":"2025-11-28T15:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:37 crc kubenswrapper[4805]: I1128 15:26:37.952806 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.003977 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:37Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.019636 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.019689 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.019702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.019746 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.019761 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.042908 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.122668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.122708 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.122720 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.122737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.122750 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.224925 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.224964 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.224974 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.224990 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.225001 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.327467 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.327519 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.327535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.327557 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.327572 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.431065 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.431121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.431134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.431151 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.431163 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.438400 4805 generic.go:334] "Generic (PLEG): container finished" podID="e746e8b8-a375-4650-98fc-f6e3ba35a50d" containerID="271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874" exitCode=0 Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.438476 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" event={"ID":"e746e8b8-a375-4650-98fc-f6e3ba35a50d","Type":"ContainerDied","Data":"271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.460552 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.476283 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.490636 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.502031 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.516406 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.535530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.535579 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.535591 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.535609 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.535623 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.538031 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.555556 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.571747 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.588860 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.603487 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.625847 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.637803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.637845 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.637859 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.637876 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.637888 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.644537 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.658904 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.675414 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.689117 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:38Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.740604 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.740658 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.740673 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.740693 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.740706 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.843705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.843745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.843755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.843770 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.843785 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.859236 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.859466 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:26:46.859427442 +0000 UTC m=+33.909218783 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.947264 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.947304 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.947315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.947331 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.947342 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:38Z","lastTransitionTime":"2025-11-28T15:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.960305 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.960443 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.960487 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:38 crc kubenswrapper[4805]: I1128 15:26:38.960541 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.960618 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.960732 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.960757 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.960759 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.960778 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.960805 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.960833 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.960751 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.960769 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:46.960733592 +0000 UTC m=+34.010524953 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.961024 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:46.9609912 +0000 UTC m=+34.010782551 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.961067 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:46.961053061 +0000 UTC m=+34.010844402 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:38 crc kubenswrapper[4805]: E1128 15:26:38.961101 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:46.961089142 +0000 UTC m=+34.010880483 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.049860 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.049891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.049899 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.049919 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.049929 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.152085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.152146 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.152162 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.152191 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.152208 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.203839 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:39 crc kubenswrapper[4805]: E1128 15:26:39.204059 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.204638 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.204656 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:39 crc kubenswrapper[4805]: E1128 15:26:39.204797 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:39 crc kubenswrapper[4805]: E1128 15:26:39.205023 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.255374 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.255415 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.255426 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.255442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.255452 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.358091 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.358135 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.358146 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.358163 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.358174 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.445531 4805 generic.go:334] "Generic (PLEG): container finished" podID="e746e8b8-a375-4650-98fc-f6e3ba35a50d" containerID="77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7" exitCode=0 Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.445593 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" event={"ID":"e746e8b8-a375-4650-98fc-f6e3ba35a50d","Type":"ContainerDied","Data":"77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.450825 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.451511 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.451570 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.451590 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.460221 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.460270 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.460281 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.460299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.460313 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.474317 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.479624 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.480488 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.496445 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.509202 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.523915 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.536015 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.545821 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.556275 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.563479 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.563514 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.563523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.563537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.563606 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.566232 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.579303 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.590297 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.602028 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.612177 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.622034 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.634249 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.649634 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.665518 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.665559 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.665569 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.665583 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.665594 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.668379 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.679771 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.690219 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.703480 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.715243 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.725815 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.743543 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.755121 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.771907 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.771966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.771981 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.772003 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.772033 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.773224 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.784395 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.794526 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.806049 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.821433 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.834074 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.847255 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:39Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.874377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.874428 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.874442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.874462 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.874474 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.978238 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.978295 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.978312 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.978335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:39 crc kubenswrapper[4805]: I1128 15:26:39.978351 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:39Z","lastTransitionTime":"2025-11-28T15:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.081765 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.081814 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.081827 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.081843 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.081856 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:40Z","lastTransitionTime":"2025-11-28T15:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.184455 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.184498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.184510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.184525 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.184537 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:40Z","lastTransitionTime":"2025-11-28T15:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.287316 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.287371 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.287386 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.287410 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.287421 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:40Z","lastTransitionTime":"2025-11-28T15:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.389466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.389518 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.389537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.389560 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.389578 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:40Z","lastTransitionTime":"2025-11-28T15:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.460338 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" event={"ID":"e746e8b8-a375-4650-98fc-f6e3ba35a50d","Type":"ContainerStarted","Data":"aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.492794 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.492869 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.492893 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.492924 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.492945 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:40Z","lastTransitionTime":"2025-11-28T15:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.493399 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.524319 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.548558 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.565078 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.580352 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.594334 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.595504 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.595529 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.595540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.595557 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.595569 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:40Z","lastTransitionTime":"2025-11-28T15:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.607073 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.627687 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.640489 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.656737 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.670117 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.683772 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.693114 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.697803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.697846 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.697855 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.698206 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.698224 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:40Z","lastTransitionTime":"2025-11-28T15:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.709714 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.721624 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:40Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.801608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.801966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.802108 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.802246 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.802411 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:40Z","lastTransitionTime":"2025-11-28T15:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.904887 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.904945 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.904954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.904970 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:40 crc kubenswrapper[4805]: I1128 15:26:40.904979 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:40Z","lastTransitionTime":"2025-11-28T15:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.007846 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.007898 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.007914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.007937 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.007952 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.110713 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.110752 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.110762 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.110776 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.110785 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.203090 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.203086 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:41 crc kubenswrapper[4805]: E1128 15:26:41.203228 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.203107 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:41 crc kubenswrapper[4805]: E1128 15:26:41.203399 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:41 crc kubenswrapper[4805]: E1128 15:26:41.203445 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.213671 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.213726 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.213737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.213752 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.213764 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.315918 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.315973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.315984 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.315997 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.316005 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.418973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.419007 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.419016 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.419031 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.419040 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.522372 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.522433 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.522446 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.522469 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.522486 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.625255 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.625300 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.625308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.625327 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.625337 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.683697 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.683746 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.683758 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.683777 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.683792 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: E1128 15:26:41.699378 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:41Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.702945 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.702983 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.702994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.703012 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.703025 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: E1128 15:26:41.719457 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:41Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.724298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.724329 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.724340 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.724394 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.724415 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: E1128 15:26:41.735641 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:41Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.739123 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.739152 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.739161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.739175 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.739184 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: E1128 15:26:41.751466 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:41Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.755374 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.755411 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.755420 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.755433 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.755446 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: E1128 15:26:41.774327 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:41Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:41 crc kubenswrapper[4805]: E1128 15:26:41.774481 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.776019 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.776058 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.776068 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.776083 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.776092 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.879742 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.879803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.879820 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.879843 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.879859 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.983484 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.983542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.983554 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.983572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:41 crc kubenswrapper[4805]: I1128 15:26:41.983584 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:41Z","lastTransitionTime":"2025-11-28T15:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.086203 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.086248 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.086258 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.086274 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.086283 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:42Z","lastTransitionTime":"2025-11-28T15:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.188873 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.188919 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.188931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.188950 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.188961 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:42Z","lastTransitionTime":"2025-11-28T15:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.292281 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.292435 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.292461 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.292537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.292558 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:42Z","lastTransitionTime":"2025-11-28T15:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.396228 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.396352 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.396445 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.396475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.396502 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:42Z","lastTransitionTime":"2025-11-28T15:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.468710 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/0.log" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.472003 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86" exitCode=1 Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.472115 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.473452 4805 scope.go:117] "RemoveContainer" containerID="b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.489399 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.499152 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.499199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.499211 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.499231 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.499244 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:42Z","lastTransitionTime":"2025-11-28T15:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.512487 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.529499 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.543020 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.558272 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.582833 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\" Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540628 6073 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540650 6073 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540786 6073 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 15:26:41.540799 6073 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 15:26:41.540810 6073 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 15:26:41.540816 6073 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 15:26:41.540835 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 15:26:41.540837 6073 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 15:26:41.540851 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 15:26:41.540865 6073 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 15:26:41.540958 6073 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.602282 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.602337 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.602349 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.602382 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.602394 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:42Z","lastTransitionTime":"2025-11-28T15:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.608657 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.624290 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.638703 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.653157 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.668389 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.678398 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.690568 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.702109 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.705563 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.705594 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.705606 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.705623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.705638 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:42Z","lastTransitionTime":"2025-11-28T15:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.716875 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:42Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.808212 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.808257 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.808271 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.808296 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.808309 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:42Z","lastTransitionTime":"2025-11-28T15:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.910646 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.910874 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.910941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.911034 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:42 crc kubenswrapper[4805]: I1128 15:26:42.911094 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:42Z","lastTransitionTime":"2025-11-28T15:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.015471 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.015523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.015540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.015565 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.015582 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.118293 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.118574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.118650 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.118733 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.118806 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.203530 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.203643 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:43 crc kubenswrapper[4805]: E1128 15:26:43.203717 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:43 crc kubenswrapper[4805]: E1128 15:26:43.203789 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.204029 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:43 crc kubenswrapper[4805]: E1128 15:26:43.204175 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.221962 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.222012 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.222025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.222047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.222063 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.223423 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.236648 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.251528 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.286876 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.311257 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.324433 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.324684 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.324763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.324825 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.324889 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.329022 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.340218 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.352645 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.368115 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\" Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540628 6073 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540650 6073 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540786 6073 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 15:26:41.540799 6073 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 15:26:41.540810 6073 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 15:26:41.540816 6073 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 15:26:41.540835 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 15:26:41.540837 6073 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 15:26:41.540851 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 15:26:41.540865 6073 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 15:26:41.540958 6073 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.385637 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.399937 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.412926 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.427694 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.427720 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.427728 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.427741 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.427751 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.428492 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.441134 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.450532 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.477623 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/0.log" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.480303 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.481073 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.496126 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.518650 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.529827 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.529868 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.529879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.529900 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.529914 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.530163 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.540459 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.553716 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.576280 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\" Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540628 6073 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540650 6073 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540786 6073 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 15:26:41.540799 6073 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 15:26:41.540810 6073 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 15:26:41.540816 6073 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 15:26:41.540835 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 15:26:41.540837 6073 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 15:26:41.540851 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 15:26:41.540865 6073 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 15:26:41.540958 6073 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.592480 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.605973 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.620697 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.632801 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.633075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.633146 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.633224 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.633341 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.634784 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.644976 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.665408 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.676190 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.692997 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.706582 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.736115 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.736155 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.736163 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.736177 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.736186 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.838980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.839057 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.839080 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.839109 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.839133 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.942805 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.942886 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.942910 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.942935 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:43 crc kubenswrapper[4805]: I1128 15:26:43.942951 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:43Z","lastTransitionTime":"2025-11-28T15:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.046543 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.046586 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.046595 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.046609 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.046618 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.149419 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.149472 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.149486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.149506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.149521 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.251812 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.251856 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.251866 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.251885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.251896 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.355561 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.355653 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.355678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.355712 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.355736 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.457980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.458017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.458027 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.458040 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.458048 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.489277 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/1.log" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.490163 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/0.log" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.492763 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e" exitCode=1 Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.492804 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.492857 4805 scope.go:117] "RemoveContainer" containerID="b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.493515 4805 scope.go:117] "RemoveContainer" containerID="ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e" Nov 28 15:26:44 crc kubenswrapper[4805]: E1128 15:26:44.493678 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.508315 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.522303 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.533024 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.542577 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.554897 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.560219 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.560279 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.560296 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.560320 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.560340 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.583949 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\" Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540628 6073 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540650 6073 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540786 6073 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 15:26:41.540799 6073 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 15:26:41.540810 6073 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 15:26:41.540816 6073 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 15:26:41.540835 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 15:26:41.540837 6073 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 15:26:41.540851 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 15:26:41.540865 6073 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 15:26:41.540958 6073 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:44Z\\\",\\\"message\\\":\\\"]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-oauth-apiserver/api_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.140\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1128 15:26:43.237889 6223 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.612666 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.626418 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.639272 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.653738 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.663161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.663223 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.663244 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.663268 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.663285 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.668432 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.681728 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.697212 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.708022 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.723375 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:44Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.765838 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.765875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.765886 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.765900 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.765913 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.868335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.868412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.868425 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.868441 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.868453 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.972118 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.972185 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.972205 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.972232 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:44 crc kubenswrapper[4805]: I1128 15:26:44.972254 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:44Z","lastTransitionTime":"2025-11-28T15:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.075637 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.075702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.075720 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.075748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.075764 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:45Z","lastTransitionTime":"2025-11-28T15:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.178620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.178654 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.178663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.178676 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.178684 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:45Z","lastTransitionTime":"2025-11-28T15:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.203555 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.203570 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:45 crc kubenswrapper[4805]: E1128 15:26:45.203712 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.203588 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:45 crc kubenswrapper[4805]: E1128 15:26:45.203844 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:45 crc kubenswrapper[4805]: E1128 15:26:45.203907 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.271225 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v"] Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.271760 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.274475 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.276567 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.281596 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.281652 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.281670 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.281696 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.281716 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:45Z","lastTransitionTime":"2025-11-28T15:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.290808 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.308178 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.324225 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2pg6\" (UniqueName: \"kubernetes.io/projected/7688879f-c48b-4b0b-9aa5-2bde152cf34b-kube-api-access-q2pg6\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.324303 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7688879f-c48b-4b0b-9aa5-2bde152cf34b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.324208 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.324353 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7688879f-c48b-4b0b-9aa5-2bde152cf34b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.324562 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7688879f-c48b-4b0b-9aa5-2bde152cf34b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.337767 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.351266 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.370097 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.384026 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.384316 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.384336 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.384348 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.384379 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.384392 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:45Z","lastTransitionTime":"2025-11-28T15:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.399121 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.410627 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.420957 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.425656 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7688879f-c48b-4b0b-9aa5-2bde152cf34b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.425701 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7688879f-c48b-4b0b-9aa5-2bde152cf34b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.425731 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2pg6\" (UniqueName: \"kubernetes.io/projected/7688879f-c48b-4b0b-9aa5-2bde152cf34b-kube-api-access-q2pg6\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.425777 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7688879f-c48b-4b0b-9aa5-2bde152cf34b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.426790 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7688879f-c48b-4b0b-9aa5-2bde152cf34b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.426890 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7688879f-c48b-4b0b-9aa5-2bde152cf34b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.430975 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7688879f-c48b-4b0b-9aa5-2bde152cf34b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.433941 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.441630 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2pg6\" (UniqueName: \"kubernetes.io/projected/7688879f-c48b-4b0b-9aa5-2bde152cf34b-kube-api-access-q2pg6\") pod \"ovnkube-control-plane-749d76644c-64s9v\" (UID: \"7688879f-c48b-4b0b-9aa5-2bde152cf34b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.445892 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.455429 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.465774 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.484536 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b03b1919d9ab8bac58347b23f730ba667fd1e5c3d5bb75efcb038234bfa2df86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:41Z\\\",\\\"message\\\":\\\" Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540628 6073 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540650 6073 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 15:26:41.540786 6073 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 15:26:41.540799 6073 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 15:26:41.540810 6073 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 15:26:41.540816 6073 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 15:26:41.540835 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 15:26:41.540837 6073 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 15:26:41.540851 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 15:26:41.540865 6073 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 15:26:41.540958 6073 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:44Z\\\",\\\"message\\\":\\\"]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-oauth-apiserver/api_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.140\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1128 15:26:43.237889 6223 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.486744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.486788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.486802 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.486819 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.486844 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:45Z","lastTransitionTime":"2025-11-28T15:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.496764 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/1.log" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.500146 4805 scope.go:117] "RemoveContainer" containerID="ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e" Nov 28 15:26:45 crc kubenswrapper[4805]: E1128 15:26:45.500296 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.500467 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.521841 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.534446 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.546973 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.560389 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.571857 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.581210 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.592584 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.592629 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.592639 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.592657 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.592668 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:45Z","lastTransitionTime":"2025-11-28T15:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.593935 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.599344 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: W1128 15:26:45.610949 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7688879f_c48b_4b0b_9aa5_2bde152cf34b.slice/crio-9aa174c270575bc38d245132d0b0d0c5f00780aab59f38342e40e34ee7ae1bb0 WatchSource:0}: Error finding container 9aa174c270575bc38d245132d0b0d0c5f00780aab59f38342e40e34ee7ae1bb0: Status 404 returned error can't find the container with id 9aa174c270575bc38d245132d0b0d0c5f00780aab59f38342e40e34ee7ae1bb0 Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.613962 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.626190 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.640150 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.651117 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.661657 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.671686 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.681919 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.692368 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.695326 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.695385 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.695406 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.695420 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.695430 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:45Z","lastTransitionTime":"2025-11-28T15:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.709970 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:44Z\\\",\\\"message\\\":\\\"]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-oauth-apiserver/api_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.140\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1128 15:26:43.237889 6223 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:45Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.798041 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.798091 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.798110 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.798130 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.798143 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:45Z","lastTransitionTime":"2025-11-28T15:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.901834 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.901883 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.901893 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.901910 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:45 crc kubenswrapper[4805]: I1128 15:26:45.901922 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:45Z","lastTransitionTime":"2025-11-28T15:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.006150 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.006217 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.006235 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.006262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.006279 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.109275 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.109323 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.109336 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.109377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.109397 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.213147 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.213208 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.213225 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.213250 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.213267 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.317212 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.317283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.317305 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.317335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.317389 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.397011 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-fplc8"] Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.397567 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:46 crc kubenswrapper[4805]: E1128 15:26:46.397643 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.420587 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.420642 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.420660 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.420685 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.420703 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.422347 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.445577 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.464461 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.479698 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.496413 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.503678 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" event={"ID":"7688879f-c48b-4b0b-9aa5-2bde152cf34b","Type":"ContainerStarted","Data":"9aa174c270575bc38d245132d0b0d0c5f00780aab59f38342e40e34ee7ae1bb0"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.516995 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.523873 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.523915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.523933 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.523957 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.523973 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.537780 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw4wh\" (UniqueName: \"kubernetes.io/projected/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-kube-api-access-dw4wh\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.537950 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.556079 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:44Z\\\",\\\"message\\\":\\\"]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-oauth-apiserver/api_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.140\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1128 15:26:43.237889 6223 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.579338 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.598296 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.618247 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.626792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.626840 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.626869 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.626894 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.626911 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.637819 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.638519 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.638558 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw4wh\" (UniqueName: \"kubernetes.io/projected/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-kube-api-access-dw4wh\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:46 crc kubenswrapper[4805]: E1128 15:26:46.638699 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:46 crc kubenswrapper[4805]: E1128 15:26:46.638768 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs podName:18cf186f-76f6-47bc-8db9-c9b7be3aaf09 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:47.138744578 +0000 UTC m=+34.188535899 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs") pod "network-metrics-daemon-fplc8" (UID: "18cf186f-76f6-47bc-8db9-c9b7be3aaf09") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.658572 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.715875 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.716800 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw4wh\" (UniqueName: \"kubernetes.io/projected/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-kube-api-access-dw4wh\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.726815 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.730755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.730792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.730803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.730820 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.730831 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.739058 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.747446 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.758575 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:46Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.833640 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.833686 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.833700 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.833721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.833736 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.936845 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.936913 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.936936 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.936967 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.936992 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:46Z","lastTransitionTime":"2025-11-28T15:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:46 crc kubenswrapper[4805]: I1128 15:26:46.941543 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:26:46 crc kubenswrapper[4805]: E1128 15:26:46.941866 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:27:02.941840567 +0000 UTC m=+49.991631918 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.040456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.040505 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.040515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.040534 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.040546 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.042406 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.042495 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.042542 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.042576 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042638 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042652 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042681 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042696 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:03.042677805 +0000 UTC m=+50.092469116 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042701 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042727 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042758 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:03.042738526 +0000 UTC m=+50.092529877 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042798 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:03.042777427 +0000 UTC m=+50.092568768 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042840 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042858 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042874 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.042917 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:03.042904161 +0000 UTC m=+50.092695502 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.143100 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.143390 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.143481 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs podName:18cf186f-76f6-47bc-8db9-c9b7be3aaf09 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:48.14345044 +0000 UTC m=+35.193241791 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs") pod "network-metrics-daemon-fplc8" (UID: "18cf186f-76f6-47bc-8db9-c9b7be3aaf09") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.144442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.144515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.144542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.144576 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.144600 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.203311 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.203436 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.203493 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.203652 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.203761 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:47 crc kubenswrapper[4805]: E1128 15:26:47.203899 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.247084 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.247114 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.247122 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.247134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.247143 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.350308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.350402 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.350425 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.350454 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.350477 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.453261 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.453318 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.453336 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.453399 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.453424 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.556085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.556146 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.556164 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.556186 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.556203 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.660205 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.660286 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.660310 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.660342 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.660398 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.763649 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.763700 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.763719 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.763744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.763761 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.866897 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.867155 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.867255 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.867342 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.867428 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.972503 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.972566 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.972580 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.972601 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:47 crc kubenswrapper[4805]: I1128 15:26:47.972620 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:47Z","lastTransitionTime":"2025-11-28T15:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.075275 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.075329 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.075341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.075376 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.075389 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:48Z","lastTransitionTime":"2025-11-28T15:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.154492 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:48 crc kubenswrapper[4805]: E1128 15:26:48.154657 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:48 crc kubenswrapper[4805]: E1128 15:26:48.154718 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs podName:18cf186f-76f6-47bc-8db9-c9b7be3aaf09 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:50.154700548 +0000 UTC m=+37.204491869 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs") pod "network-metrics-daemon-fplc8" (UID: "18cf186f-76f6-47bc-8db9-c9b7be3aaf09") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.177797 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.177837 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.177846 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.177861 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.177870 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:48Z","lastTransitionTime":"2025-11-28T15:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.203617 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:48 crc kubenswrapper[4805]: E1128 15:26:48.203781 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.280996 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.281034 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.281043 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.281058 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.281072 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:48Z","lastTransitionTime":"2025-11-28T15:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.384627 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.384685 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.384703 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.384729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.384749 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:48Z","lastTransitionTime":"2025-11-28T15:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.488562 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.488635 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.488660 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.488691 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.488714 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:48Z","lastTransitionTime":"2025-11-28T15:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.514030 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" event={"ID":"7688879f-c48b-4b0b-9aa5-2bde152cf34b","Type":"ContainerStarted","Data":"a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.514131 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" event={"ID":"7688879f-c48b-4b0b-9aa5-2bde152cf34b","Type":"ContainerStarted","Data":"2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.539331 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.563278 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:44Z\\\",\\\"message\\\":\\\"]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-oauth-apiserver/api_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.140\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1128 15:26:43.237889 6223 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.582004 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.591304 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.591623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.591746 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.591868 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.591989 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:48Z","lastTransitionTime":"2025-11-28T15:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.607915 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.623833 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.637133 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.654238 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.670991 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.684747 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.694624 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.694680 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.694697 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.694721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.694737 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:48Z","lastTransitionTime":"2025-11-28T15:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.701634 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.715411 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.737599 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.757413 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.771710 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.782519 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.790728 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.796594 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.796625 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.796633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.796648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.796657 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:48Z","lastTransitionTime":"2025-11-28T15:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.800427 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:48Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.899941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.900312 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.900647 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.901007 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:48 crc kubenswrapper[4805]: I1128 15:26:48.901214 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:48Z","lastTransitionTime":"2025-11-28T15:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.004778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.005041 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.005155 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.005251 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.005349 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.108914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.108961 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.108973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.108990 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.109002 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.203080 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:49 crc kubenswrapper[4805]: E1128 15:26:49.203859 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.203204 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.203105 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:49 crc kubenswrapper[4805]: E1128 15:26:49.204500 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:49 crc kubenswrapper[4805]: E1128 15:26:49.204413 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.211703 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.211996 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.212118 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.212205 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.212278 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.315379 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.315412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.315420 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.315434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.315443 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.418514 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.418866 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.419014 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.419155 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.419288 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.521851 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.522134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.522283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.522419 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.522543 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.625868 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.625930 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.625951 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.625973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.625988 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.728503 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.728570 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.728589 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.728650 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.728669 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.831213 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.831282 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.831298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.831316 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.831327 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.934471 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.934514 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.934527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.934546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:49 crc kubenswrapper[4805]: I1128 15:26:49.934557 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:49Z","lastTransitionTime":"2025-11-28T15:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.037715 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.037769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.037782 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.037800 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.037814 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.140880 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.141227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.141638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.142024 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.142413 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.181145 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:50 crc kubenswrapper[4805]: E1128 15:26:50.181538 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:50 crc kubenswrapper[4805]: E1128 15:26:50.181828 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs podName:18cf186f-76f6-47bc-8db9-c9b7be3aaf09 nodeName:}" failed. No retries permitted until 2025-11-28 15:26:54.181801585 +0000 UTC m=+41.231592936 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs") pod "network-metrics-daemon-fplc8" (UID: "18cf186f-76f6-47bc-8db9-c9b7be3aaf09") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.203392 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:50 crc kubenswrapper[4805]: E1128 15:26:50.204352 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.245987 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.246085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.246112 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.246143 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.246166 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.349072 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.349479 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.349619 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.349970 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.350104 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.452480 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.452508 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.452516 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.452530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.452539 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.555380 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.555637 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.555716 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.555824 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.555903 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.659074 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.659179 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.659214 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.659239 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.659256 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.763275 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.763337 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.763385 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.763417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.763441 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.774933 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.797526 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.827520 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:44Z\\\",\\\"message\\\":\\\"]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-oauth-apiserver/api_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.140\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1128 15:26:43.237889 6223 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.862758 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.867437 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.867493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.867516 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.867542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.867561 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.887109 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.906850 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.927015 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.948587 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.964388 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.969792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.969835 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.969867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.969885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.969898 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:50Z","lastTransitionTime":"2025-11-28T15:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.975220 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:50 crc kubenswrapper[4805]: I1128 15:26:50.991501 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:50Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.006771 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:51Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.020838 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:51Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.037124 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:51Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.053674 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:51Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.071296 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:51Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.072589 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.072656 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.072679 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.072710 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.072732 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:51Z","lastTransitionTime":"2025-11-28T15:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.081514 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:51Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.091474 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:51Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.175894 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.175981 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.176054 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.176128 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.176154 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:51Z","lastTransitionTime":"2025-11-28T15:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.203517 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:51 crc kubenswrapper[4805]: E1128 15:26:51.203699 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.204200 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:51 crc kubenswrapper[4805]: E1128 15:26:51.204339 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.204486 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:51 crc kubenswrapper[4805]: E1128 15:26:51.204722 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.278857 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.278910 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.278922 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.278938 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.278949 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:51Z","lastTransitionTime":"2025-11-28T15:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.381956 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.382273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.382377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.382470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.382589 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:51Z","lastTransitionTime":"2025-11-28T15:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.485955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.486027 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.486042 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.486063 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.486077 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:51Z","lastTransitionTime":"2025-11-28T15:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.590598 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.590641 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.590650 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.590664 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.590675 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:51Z","lastTransitionTime":"2025-11-28T15:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.694429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.694512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.694535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.694567 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.694589 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:51Z","lastTransitionTime":"2025-11-28T15:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.798096 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.798158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.798172 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.798194 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.798215 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:51Z","lastTransitionTime":"2025-11-28T15:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.901257 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.901334 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.901352 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.901411 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:51 crc kubenswrapper[4805]: I1128 15:26:51.901431 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:51Z","lastTransitionTime":"2025-11-28T15:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.004201 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.004474 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.004505 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.004536 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.004563 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.093204 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.093315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.093336 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.093396 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.093423 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: E1128 15:26:52.113409 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:52Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.118955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.119227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.119408 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.119512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.119587 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: E1128 15:26:52.137802 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:52Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.144395 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.144474 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.144493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.144525 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.144544 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: E1128 15:26:52.166758 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:52Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.173005 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.173140 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.173219 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.173321 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.173434 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: E1128 15:26:52.190676 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:52Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.195530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.195705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.195809 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.195909 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.196001 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.203238 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:52 crc kubenswrapper[4805]: E1128 15:26:52.203456 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:26:52 crc kubenswrapper[4805]: E1128 15:26:52.210584 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:52Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:52 crc kubenswrapper[4805]: E1128 15:26:52.211637 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.213993 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.214043 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.214057 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.214075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.214088 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.317102 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.317149 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.317161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.317177 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.317188 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.420969 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.421015 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.421027 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.421044 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.421058 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.524299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.524407 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.524423 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.524447 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.524464 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.627545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.627612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.627630 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.627655 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.627674 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.730499 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.730549 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.730565 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.730588 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.730606 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.833196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.833279 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.833303 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.833328 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.833347 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.937052 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.937117 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.937133 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.937159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:52 crc kubenswrapper[4805]: I1128 15:26:52.937177 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:52Z","lastTransitionTime":"2025-11-28T15:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.039959 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.040289 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.040350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.041517 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.041952 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.144982 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.145023 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.145035 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.145051 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.145063 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.203268 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.203286 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.203584 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:53 crc kubenswrapper[4805]: E1128 15:26:53.203503 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:53 crc kubenswrapper[4805]: E1128 15:26:53.203853 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:53 crc kubenswrapper[4805]: E1128 15:26:53.204052 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.229046 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.248220 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.248292 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.248310 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.248333 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.248353 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.249313 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.274693 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.292327 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.305817 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.319763 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.333673 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.349418 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.351831 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.352228 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.352276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.352308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.352331 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.362932 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.385791 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:44Z\\\",\\\"message\\\":\\\"]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-oauth-apiserver/api_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.140\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1128 15:26:43.237889 6223 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.407117 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.419442 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.435615 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.455407 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.455457 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.455470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.455489 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.455502 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.467275 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.484611 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.500161 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.515458 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:53Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.559405 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.559509 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.559530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.559558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.559576 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.662485 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.662533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.662546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.662562 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.662573 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.765051 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.765117 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.765138 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.765165 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.765182 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.867637 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.867705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.867726 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.867756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.867776 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.970602 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.970652 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.970662 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.970679 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:53 crc kubenswrapper[4805]: I1128 15:26:53.970688 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:53Z","lastTransitionTime":"2025-11-28T15:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.073776 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.073850 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.073890 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.073917 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.073936 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:54Z","lastTransitionTime":"2025-11-28T15:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.176758 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.176815 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.176832 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.176856 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.176873 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:54Z","lastTransitionTime":"2025-11-28T15:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.203748 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:54 crc kubenswrapper[4805]: E1128 15:26:54.203979 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.227260 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:54 crc kubenswrapper[4805]: E1128 15:26:54.227505 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:54 crc kubenswrapper[4805]: E1128 15:26:54.227629 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs podName:18cf186f-76f6-47bc-8db9-c9b7be3aaf09 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:02.227601281 +0000 UTC m=+49.277392622 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs") pod "network-metrics-daemon-fplc8" (UID: "18cf186f-76f6-47bc-8db9-c9b7be3aaf09") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.280378 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.280416 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.280427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.280442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.280454 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:54Z","lastTransitionTime":"2025-11-28T15:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.383215 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.383264 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.383275 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.383294 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.383303 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:54Z","lastTransitionTime":"2025-11-28T15:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.485865 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.485909 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.485921 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.485936 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.485952 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:54Z","lastTransitionTime":"2025-11-28T15:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.590159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.590247 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.590265 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.590294 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.590312 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:54Z","lastTransitionTime":"2025-11-28T15:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.694095 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.694428 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.694539 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.694631 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.694718 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:54Z","lastTransitionTime":"2025-11-28T15:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.797651 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.797716 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.797738 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.797763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.797780 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:54Z","lastTransitionTime":"2025-11-28T15:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.900060 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.900437 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.900617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.900768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:54 crc kubenswrapper[4805]: I1128 15:26:54.900897 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:54Z","lastTransitionTime":"2025-11-28T15:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.005256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.005329 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.005404 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.005446 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.005469 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.108159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.108219 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.108235 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.108259 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.108275 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.203868 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.203892 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:55 crc kubenswrapper[4805]: E1128 15:26:55.204520 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:55 crc kubenswrapper[4805]: E1128 15:26:55.204520 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.203992 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:55 crc kubenswrapper[4805]: E1128 15:26:55.205051 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.211078 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.211131 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.211147 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.211171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.211189 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.313620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.313671 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.313682 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.313702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.313714 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.417232 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.417299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.417311 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.417335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.417349 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.520297 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.520431 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.520445 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.520466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.520477 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.623390 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.623420 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.623429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.623444 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.623453 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.725957 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.726017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.726040 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.726068 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.726089 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.829232 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.829304 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.829332 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.829401 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.829427 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.932884 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.932965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.932997 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.933027 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:55 crc kubenswrapper[4805]: I1128 15:26:55.933051 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:55Z","lastTransitionTime":"2025-11-28T15:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.036467 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.036533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.036550 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.036575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.036594 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.139163 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.139208 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.139219 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.139238 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.139252 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.203683 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:56 crc kubenswrapper[4805]: E1128 15:26:56.203819 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.241774 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.242077 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.242228 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.242403 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.242535 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.344597 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.344649 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.344661 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.344675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.344701 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.447652 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.447761 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.447786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.447821 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.447849 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.551106 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.551511 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.551631 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.551735 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.551805 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.655152 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.655241 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.655267 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.655330 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.655350 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.759270 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.759726 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.759851 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.759992 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.760118 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.863042 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.863098 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.863113 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.863137 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.863159 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.966218 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.966290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.966305 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.966325 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:56 crc kubenswrapper[4805]: I1128 15:26:56.966339 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:56Z","lastTransitionTime":"2025-11-28T15:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.071199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.071288 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.071338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.071407 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.071434 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:57Z","lastTransitionTime":"2025-11-28T15:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.174984 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.175054 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.175073 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.175096 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.175113 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:57Z","lastTransitionTime":"2025-11-28T15:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.203735 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.203882 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.204061 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:57 crc kubenswrapper[4805]: E1128 15:26:57.204059 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:57 crc kubenswrapper[4805]: E1128 15:26:57.204177 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:57 crc kubenswrapper[4805]: E1128 15:26:57.204277 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.278406 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.278461 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.278470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.278517 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.278529 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:57Z","lastTransitionTime":"2025-11-28T15:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.381668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.381716 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.381728 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.381749 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.381762 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:57Z","lastTransitionTime":"2025-11-28T15:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.485598 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.485662 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.485673 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.485692 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.485703 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:57Z","lastTransitionTime":"2025-11-28T15:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.589230 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.589321 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.589342 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.589406 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.589427 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:57Z","lastTransitionTime":"2025-11-28T15:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.693300 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.693383 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.693400 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.693419 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.693429 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:57Z","lastTransitionTime":"2025-11-28T15:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.797032 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.797075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.797085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.797103 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.797115 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:57Z","lastTransitionTime":"2025-11-28T15:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.900407 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.900450 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.900459 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.900476 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:57 crc kubenswrapper[4805]: I1128 15:26:57.900486 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:57Z","lastTransitionTime":"2025-11-28T15:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.003575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.003613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.003623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.003641 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.003651 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.106666 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.106708 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.106735 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.106757 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.106769 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.203410 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:26:58 crc kubenswrapper[4805]: E1128 15:26:58.203623 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.210079 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.210130 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.210142 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.210160 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.210174 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.313439 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.313498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.313509 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.313531 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.313545 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.416940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.416983 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.416994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.417010 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.417023 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.520097 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.520187 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.520218 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.520246 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.520263 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.623314 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.623437 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.623465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.623497 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.623520 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.726725 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.726787 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.726820 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.726845 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.726861 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.830156 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.830225 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.830246 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.830290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.830308 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.933407 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.933496 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.933515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.933541 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:58 crc kubenswrapper[4805]: I1128 15:26:58.933559 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:58Z","lastTransitionTime":"2025-11-28T15:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.036458 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.036505 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.036517 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.036533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.036545 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.143295 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.143384 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.143404 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.143429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.143447 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.203704 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.203807 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:26:59 crc kubenswrapper[4805]: E1128 15:26:59.203906 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:26:59 crc kubenswrapper[4805]: E1128 15:26:59.204028 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.204138 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:26:59 crc kubenswrapper[4805]: E1128 15:26:59.204228 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.205805 4805 scope.go:117] "RemoveContainer" containerID="ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.260432 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.260470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.260482 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.260500 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.260513 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.363644 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.363684 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.363695 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.363712 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.363724 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.466074 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.466115 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.466126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.466144 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.466156 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.556548 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/1.log" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.559945 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.560440 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.568854 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.568899 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.568911 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.568929 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.568941 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.582065 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.599110 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.619426 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.637378 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.648979 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.664046 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.670908 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.670958 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.670973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.670996 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.671010 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.694336 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:44Z\\\",\\\"message\\\":\\\"]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-oauth-apiserver/api_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.140\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1128 15:26:43.237889 6223 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.713312 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.734034 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.744669 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.754899 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.774290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.774344 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.774373 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.774392 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.774405 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.774920 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.790306 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.805011 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.817872 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.829646 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.849205 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:26:59Z is after 2025-08-24T17:21:41Z" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.876758 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.876799 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.876809 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.876823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.876833 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.980035 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.980337 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.980345 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.980375 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:26:59 crc kubenswrapper[4805]: I1128 15:26:59.980391 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:26:59Z","lastTransitionTime":"2025-11-28T15:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.089072 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.089108 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.089120 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.089136 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.089148 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:00Z","lastTransitionTime":"2025-11-28T15:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.192237 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.192290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.192300 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.192317 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.192328 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:00Z","lastTransitionTime":"2025-11-28T15:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.203725 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:00 crc kubenswrapper[4805]: E1128 15:27:00.203923 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.295551 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.295606 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.295616 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.295633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.295645 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:00Z","lastTransitionTime":"2025-11-28T15:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.398049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.398416 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.398434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.398452 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.398464 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:00Z","lastTransitionTime":"2025-11-28T15:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.501306 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.501390 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.501408 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.501432 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.501448 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:00Z","lastTransitionTime":"2025-11-28T15:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.567188 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/2.log" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.568225 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/1.log" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.577609 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae" exitCode=1 Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.577681 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.577739 4805 scope.go:117] "RemoveContainer" containerID="ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.578854 4805 scope.go:117] "RemoveContainer" containerID="ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae" Nov 28 15:27:00 crc kubenswrapper[4805]: E1128 15:27:00.579163 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.603807 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.603904 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.603915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.603931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.603941 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:00Z","lastTransitionTime":"2025-11-28T15:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.618427 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.639130 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.657621 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.675673 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.693667 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.706524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.706575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.706592 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.706611 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.706626 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:00Z","lastTransitionTime":"2025-11-28T15:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.707156 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.719508 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.736152 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.747215 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.761052 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.773125 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.785994 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.797477 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.807196 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.808708 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.808743 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.808758 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.808781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.808797 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:00Z","lastTransitionTime":"2025-11-28T15:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.816640 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.827511 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.847058 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef04a58c647c5268266e7c428ca94642b18bd5cdd410a18dfdb23739b32c433e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:26:44Z\\\",\\\"message\\\":\\\"]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-oauth-apiserver/api_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.140\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1128 15:26:43.237889 6223 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:00Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.911672 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.911740 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.911764 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.911795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:00 crc kubenswrapper[4805]: I1128 15:27:00.911817 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:00Z","lastTransitionTime":"2025-11-28T15:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.014885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.014935 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.014947 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.014975 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.014986 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.117738 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.117841 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.117882 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.117923 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.117934 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.203598 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.203634 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:01 crc kubenswrapper[4805]: E1128 15:27:01.203813 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.203834 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:01 crc kubenswrapper[4805]: E1128 15:27:01.204639 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:01 crc kubenswrapper[4805]: E1128 15:27:01.204897 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.220773 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.220842 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.220864 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.220891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.220913 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.323464 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.323519 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.323535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.323558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.323576 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.425995 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.426074 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.426097 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.426129 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.426152 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.528805 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.528856 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.528867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.528886 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.528898 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.584462 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/2.log" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.589670 4805 scope.go:117] "RemoveContainer" containerID="ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae" Nov 28 15:27:01 crc kubenswrapper[4805]: E1128 15:27:01.589950 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.608025 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.624848 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.632071 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.632134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.632153 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.632179 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.632196 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.642492 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.654144 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.668426 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.683843 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.699138 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.716272 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.734410 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.736447 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.736514 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.736529 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.736553 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.736570 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.758441 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.776345 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.787732 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.804304 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.835432 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.839427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.839464 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.839475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.839491 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.839509 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.862303 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.876404 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.898090 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:01Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.942904 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.942992 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.943014 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.943078 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:01 crc kubenswrapper[4805]: I1128 15:27:01.943102 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:01Z","lastTransitionTime":"2025-11-28T15:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.046695 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.046782 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.046802 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.046831 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.046850 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.149882 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.150244 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.150527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.150723 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.150889 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.203596 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:02 crc kubenswrapper[4805]: E1128 15:27:02.203764 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.253135 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.253180 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.253189 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.253206 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.253219 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.317422 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:02 crc kubenswrapper[4805]: E1128 15:27:02.317634 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:27:02 crc kubenswrapper[4805]: E1128 15:27:02.317740 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs podName:18cf186f-76f6-47bc-8db9-c9b7be3aaf09 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:18.317711469 +0000 UTC m=+65.367502810 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs") pod "network-metrics-daemon-fplc8" (UID: "18cf186f-76f6-47bc-8db9-c9b7be3aaf09") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.355803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.355845 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.355856 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.355874 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.355887 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.458766 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.458810 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.458823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.458842 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.458856 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.498602 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.499100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.499294 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.499548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.499696 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: E1128 15:27:02.520509 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:02Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.525793 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.526061 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.526256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.526530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.526705 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: E1128 15:27:02.548657 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:02Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.554475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.554537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.554557 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.554584 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.554604 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: E1128 15:27:02.574565 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:02Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.579803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.579843 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.579857 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.579876 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.579888 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: E1128 15:27:02.604095 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:02Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.609305 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.609507 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.609572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.609638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.609702 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: E1128 15:27:02.625931 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:02Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:02 crc kubenswrapper[4805]: E1128 15:27:02.626318 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.628745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.628870 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.628933 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.629010 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.629081 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.731888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.731941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.731955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.731975 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.731988 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.836318 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.836419 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.836437 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.836463 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.836486 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.939926 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.940022 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.940045 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.940078 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:02 crc kubenswrapper[4805]: I1128 15:27:02.940105 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:02Z","lastTransitionTime":"2025-11-28T15:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.028604 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.028986 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:27:35.028963007 +0000 UTC m=+82.078754318 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.043577 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.043613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.043622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.043664 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.043674 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.130100 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.130172 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.130210 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.130266 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130392 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130481 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130508 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130522 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130542 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130573 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130522 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:35.130492505 +0000 UTC m=+82.180283896 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130532 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130643 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130648 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:35.130627569 +0000 UTC m=+82.180418880 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130667 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:35.13066038 +0000 UTC m=+82.180451691 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.130679 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:35.13067364 +0000 UTC m=+82.180464941 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.146828 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.147197 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.147392 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.147539 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.147721 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.203655 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.204082 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.203872 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.204274 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.203748 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:03 crc kubenswrapper[4805]: E1128 15:27:03.204496 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.224043 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.243561 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.250959 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.251031 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.251053 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.251083 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.251106 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.269248 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.293264 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.312352 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.331664 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.346650 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.354612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.354673 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.354700 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.354730 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.354753 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.360806 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.382060 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.405076 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.438292 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.457619 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.457918 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.458059 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.458248 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.458444 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.460117 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.481322 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.500110 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.524705 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.542060 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.560760 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:03Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.562921 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.562998 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.563038 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.563098 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.563128 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.667176 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.667235 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.667250 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.667272 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.667286 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.771027 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.771472 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.771639 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.771796 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.771948 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.875037 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.875070 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.875078 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.875111 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.875133 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.978350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.978427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.978442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.978464 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:03 crc kubenswrapper[4805]: I1128 15:27:03.978479 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:03Z","lastTransitionTime":"2025-11-28T15:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.080971 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.081016 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.081027 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.081044 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.081056 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:04Z","lastTransitionTime":"2025-11-28T15:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.183577 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.183642 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.183662 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.183705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.183733 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:04Z","lastTransitionTime":"2025-11-28T15:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.203662 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:04 crc kubenswrapper[4805]: E1128 15:27:04.204209 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.287416 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.287472 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.287487 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.287510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.287523 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:04Z","lastTransitionTime":"2025-11-28T15:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.390455 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.390507 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.390517 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.390535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.390546 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:04Z","lastTransitionTime":"2025-11-28T15:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.493768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.493830 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.493843 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.493867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.493880 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:04Z","lastTransitionTime":"2025-11-28T15:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.597503 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.597549 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.597558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.597574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.597584 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:04Z","lastTransitionTime":"2025-11-28T15:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.701265 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.701303 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.701315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.701333 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.701344 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:04Z","lastTransitionTime":"2025-11-28T15:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.804852 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.804922 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.804940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.804967 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.804985 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:04Z","lastTransitionTime":"2025-11-28T15:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.907710 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.907768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.907786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.907809 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:04 crc kubenswrapper[4805]: I1128 15:27:04.907826 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:04Z","lastTransitionTime":"2025-11-28T15:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.011675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.011716 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.011730 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.011749 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.011761 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.114895 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.114939 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.114949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.114970 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.114983 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.203750 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:05 crc kubenswrapper[4805]: E1128 15:27:05.203937 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.204324 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:05 crc kubenswrapper[4805]: E1128 15:27:05.204525 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.204861 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:05 crc kubenswrapper[4805]: E1128 15:27:05.205003 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.217208 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.217257 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.217279 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.217301 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.217316 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.320171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.320240 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.320258 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.320283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.320303 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.423855 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.423910 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.423931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.423955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.423969 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.527554 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.527629 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.527656 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.527691 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.527716 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.631239 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.631338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.631391 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.631417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.631435 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.735740 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.735823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.735846 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.735893 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.735914 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.839109 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.839173 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.839191 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.839216 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.839248 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.942664 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.942724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.942761 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.942791 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:05 crc kubenswrapper[4805]: I1128 15:27:05.942814 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:05Z","lastTransitionTime":"2025-11-28T15:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.045940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.046025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.046044 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.046069 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.046087 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.149645 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.150086 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.150250 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.150434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.150590 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.203957 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:06 crc kubenswrapper[4805]: E1128 15:27:06.204190 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.254661 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.254979 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.255093 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.255211 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.255306 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.358522 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.358914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.359061 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.359196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.359325 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.462509 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.462575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.462592 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.462617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.462637 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.565864 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.565917 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.565934 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.565957 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.565974 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.668481 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.668556 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.668574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.668608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.668628 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.772231 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.772310 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.772329 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.772383 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.772403 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.875326 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.875381 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.875390 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.875405 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.875414 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.978833 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.978881 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.978890 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.978906 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:06 crc kubenswrapper[4805]: I1128 15:27:06.978915 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:06Z","lastTransitionTime":"2025-11-28T15:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.082329 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.082442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.082484 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.082520 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.082550 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:07Z","lastTransitionTime":"2025-11-28T15:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.185472 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.185523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.185538 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.185556 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.185568 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:07Z","lastTransitionTime":"2025-11-28T15:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.203004 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:07 crc kubenswrapper[4805]: E1128 15:27:07.203140 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.203173 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.203190 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:07 crc kubenswrapper[4805]: E1128 15:27:07.203338 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:07 crc kubenswrapper[4805]: E1128 15:27:07.203567 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.288115 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.288185 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.288209 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.288237 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.288256 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:07Z","lastTransitionTime":"2025-11-28T15:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.391869 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.391931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.391954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.391984 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.392008 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:07Z","lastTransitionTime":"2025-11-28T15:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.494737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.494798 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.494808 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.494823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.494834 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:07Z","lastTransitionTime":"2025-11-28T15:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.597406 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.597470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.597487 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.597519 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.597539 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:07Z","lastTransitionTime":"2025-11-28T15:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.699763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.699820 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.699838 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.699862 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.699879 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:07Z","lastTransitionTime":"2025-11-28T15:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.802822 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.802890 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.802912 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.802941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.802985 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:07Z","lastTransitionTime":"2025-11-28T15:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.906424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.906465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.906475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.906493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:07 crc kubenswrapper[4805]: I1128 15:27:07.906505 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:07Z","lastTransitionTime":"2025-11-28T15:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.011182 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.011257 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.011276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.011367 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.011482 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.050234 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.064549 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.073983 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.094246 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.114284 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.115211 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.115290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.115306 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.115324 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.115337 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.124387 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.136649 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.154790 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.176796 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.195730 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.203847 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:08 crc kubenswrapper[4805]: E1128 15:27:08.203945 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.210694 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.218306 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.218342 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.218354 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.218389 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.218401 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.224931 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.236246 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.246171 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.259726 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.284971 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.302613 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.320673 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.320722 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.320735 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.320753 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.320765 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.320981 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.337476 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:08Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.423449 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.423524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.423548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.423580 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.423606 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.526776 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.526861 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.526892 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.526926 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.526954 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.629758 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.629817 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.629834 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.629857 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.629873 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.732344 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.732419 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.732430 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.732448 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.732459 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.836047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.836127 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.836151 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.836184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.836207 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.939282 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.939368 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.939412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.939438 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:08 crc kubenswrapper[4805]: I1128 15:27:08.939458 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:08Z","lastTransitionTime":"2025-11-28T15:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.042574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.042637 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.042648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.042663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.042672 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.145484 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.145566 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.145581 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.145604 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.145619 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.203542 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.203681 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:09 crc kubenswrapper[4805]: E1128 15:27:09.203938 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.203991 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:09 crc kubenswrapper[4805]: E1128 15:27:09.204093 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:09 crc kubenswrapper[4805]: E1128 15:27:09.204246 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.248328 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.248451 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.248471 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.248499 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.248519 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.351622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.351687 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.351705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.351734 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.351752 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.454683 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.454738 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.454751 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.454771 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.454787 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.558599 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.558659 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.558676 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.558702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.558723 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.661670 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.661754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.661787 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.661816 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.661838 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.764640 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.764697 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.764711 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.764729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.764743 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.867853 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.867897 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.867911 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.867928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.867941 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.971705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.971778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.971795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.971821 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:09 crc kubenswrapper[4805]: I1128 15:27:09.971839 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:09Z","lastTransitionTime":"2025-11-28T15:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.074291 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.074319 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.074352 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.074390 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.074399 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:10Z","lastTransitionTime":"2025-11-28T15:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.178118 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.178183 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.178196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.178216 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.178232 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:10Z","lastTransitionTime":"2025-11-28T15:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.203915 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:10 crc kubenswrapper[4805]: E1128 15:27:10.204203 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.281074 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.281206 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.281226 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.281251 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.281275 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:10Z","lastTransitionTime":"2025-11-28T15:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.384882 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.384947 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.384969 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.385001 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.385023 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:10Z","lastTransitionTime":"2025-11-28T15:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.488434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.488496 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.488512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.488535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.488553 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:10Z","lastTransitionTime":"2025-11-28T15:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.591940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.592034 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.592063 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.592097 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.592124 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:10Z","lastTransitionTime":"2025-11-28T15:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.695865 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.695941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.695964 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.695988 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.696009 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:10Z","lastTransitionTime":"2025-11-28T15:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.799238 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.799295 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.799309 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.799333 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.799351 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:10Z","lastTransitionTime":"2025-11-28T15:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.902728 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.902787 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.902799 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.902820 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:10 crc kubenswrapper[4805]: I1128 15:27:10.902833 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:10Z","lastTransitionTime":"2025-11-28T15:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.005690 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.005744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.005760 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.005784 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.005800 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.109467 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.109531 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.109548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.109572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.109592 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.203254 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.203466 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:11 crc kubenswrapper[4805]: E1128 15:27:11.203628 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.203673 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:11 crc kubenswrapper[4805]: E1128 15:27:11.203903 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:11 crc kubenswrapper[4805]: E1128 15:27:11.204043 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.212094 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.212142 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.212158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.212179 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.212196 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.314796 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.314873 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.314890 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.314919 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.314941 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.418293 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.418409 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.418438 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.418469 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.418491 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.521552 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.521613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.521624 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.521650 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.521663 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.632526 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.632577 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.632588 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.632607 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.632619 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.735948 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.736009 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.736026 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.736051 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.736070 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.838813 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.838870 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.838886 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.838907 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.838924 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.941949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.942011 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.942028 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.942052 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:11 crc kubenswrapper[4805]: I1128 15:27:11.942072 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:11Z","lastTransitionTime":"2025-11-28T15:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.044758 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.044793 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.044804 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.044820 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.044832 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.147723 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.147777 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.147789 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.147808 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.147820 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.203700 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:12 crc kubenswrapper[4805]: E1128 15:27:12.203958 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.251484 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.251552 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.251569 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.251595 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.251613 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.354614 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.354663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.354675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.354695 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.354708 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.457861 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.457901 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.457910 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.457927 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.457935 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.560572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.560633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.560649 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.560674 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.560690 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.664253 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.664754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.664948 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.665141 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.665569 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.757631 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.757676 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.757688 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.757708 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.757719 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: E1128 15:27:12.788511 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:12Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.794712 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.794748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.794758 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.794774 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.794785 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: E1128 15:27:12.813950 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:12Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.821879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.821918 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.821928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.821948 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.821960 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: E1128 15:27:12.838823 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:12Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.843526 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.843596 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.843614 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.843643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.843663 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: E1128 15:27:12.862071 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:12Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.868206 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.868266 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.868285 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.868308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.868326 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: E1128 15:27:12.881793 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:12Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:12 crc kubenswrapper[4805]: E1128 15:27:12.881969 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.884184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.884246 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.884259 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.884284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.884299 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.987847 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.987903 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.987915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.987941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:12 crc kubenswrapper[4805]: I1128 15:27:12.987956 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:12Z","lastTransitionTime":"2025-11-28T15:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.090964 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.091017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.091032 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.091053 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.091069 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:13Z","lastTransitionTime":"2025-11-28T15:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.193757 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.193831 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.193843 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.193861 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.193874 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:13Z","lastTransitionTime":"2025-11-28T15:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.203029 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.203094 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:13 crc kubenswrapper[4805]: E1128 15:27:13.203217 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:13 crc kubenswrapper[4805]: E1128 15:27:13.203838 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.203892 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:13 crc kubenswrapper[4805]: E1128 15:27:13.204492 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.204791 4805 scope.go:117] "RemoveContainer" containerID="ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae" Nov 28 15:27:13 crc kubenswrapper[4805]: E1128 15:27:13.205056 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.219298 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.234276 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.250814 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.268420 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.285729 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.296029 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.296091 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.296116 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.296147 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.296171 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:13Z","lastTransitionTime":"2025-11-28T15:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.306474 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"030619ba-de4f-4db2-8138-4744768d40a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.326901 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.357116 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.371683 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.387541 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.398543 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.398585 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.398601 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.398621 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.398635 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:13Z","lastTransitionTime":"2025-11-28T15:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.402303 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.427500 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.442603 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.458945 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.474504 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.487600 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.501336 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.501369 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.501413 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.501428 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.501437 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:13Z","lastTransitionTime":"2025-11-28T15:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.506009 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.520078 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:13Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.604399 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.604475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.604545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.604609 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.604625 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:13Z","lastTransitionTime":"2025-11-28T15:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.707725 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.707768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.707777 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.707792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.707803 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:13Z","lastTransitionTime":"2025-11-28T15:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.811076 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.811121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.811154 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.811173 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.811185 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:13Z","lastTransitionTime":"2025-11-28T15:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.914888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.914954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.914973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.914998 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:13 crc kubenswrapper[4805]: I1128 15:27:13.915017 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:13Z","lastTransitionTime":"2025-11-28T15:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.017981 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.018025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.018039 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.018060 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.018074 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.121757 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.121834 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.121858 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.121883 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.121904 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.203897 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:14 crc kubenswrapper[4805]: E1128 15:27:14.204116 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.225179 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.225525 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.225687 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.225827 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.225949 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.329944 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.330029 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.330055 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.330088 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.330111 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.433519 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.433562 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.433572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.433591 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.433601 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.538297 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.538350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.538424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.538452 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.538469 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.640649 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.640709 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.640726 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.640754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.640771 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.743022 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.743098 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.743110 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.743132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.743146 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.846468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.846720 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.846738 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.846763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.846781 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.951306 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.952270 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.952585 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.952826 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:14 crc kubenswrapper[4805]: I1128 15:27:14.953018 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:14Z","lastTransitionTime":"2025-11-28T15:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.055364 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.055404 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.055412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.055426 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.055434 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.158298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.158337 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.158348 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.158367 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.158402 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.203454 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:15 crc kubenswrapper[4805]: E1128 15:27:15.203591 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.203856 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:15 crc kubenswrapper[4805]: E1128 15:27:15.203927 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.204068 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:15 crc kubenswrapper[4805]: E1128 15:27:15.204155 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.261240 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.261602 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.261700 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.261817 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.261917 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.364582 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.364624 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.364633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.364648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.364658 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.467283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.467327 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.467344 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.467396 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.467415 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.569953 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.570255 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.570388 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.570503 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.570585 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.673368 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.673419 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.673431 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.673448 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.673461 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.777240 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.777767 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.778676 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.779088 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.779470 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.881726 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.881768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.881778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.881795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.881808 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.984838 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.984878 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.984890 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.984907 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:15 crc kubenswrapper[4805]: I1128 15:27:15.984919 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:15Z","lastTransitionTime":"2025-11-28T15:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.088451 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.088509 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.088520 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.088540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.088552 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:16Z","lastTransitionTime":"2025-11-28T15:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.191849 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.191926 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.191965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.191997 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.192022 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:16Z","lastTransitionTime":"2025-11-28T15:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.203605 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:16 crc kubenswrapper[4805]: E1128 15:27:16.203846 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.294675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.294709 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.294721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.294739 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.294751 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:16Z","lastTransitionTime":"2025-11-28T15:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.397229 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.397271 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.397283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.397301 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.397312 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:16Z","lastTransitionTime":"2025-11-28T15:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.499421 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.499459 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.499468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.499484 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.499492 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:16Z","lastTransitionTime":"2025-11-28T15:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.602141 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.602180 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.602188 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.602203 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.602214 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:16Z","lastTransitionTime":"2025-11-28T15:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.705224 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.705257 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.705266 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.705280 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.705291 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:16Z","lastTransitionTime":"2025-11-28T15:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.808572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.808616 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.808627 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.808643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.808656 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:16Z","lastTransitionTime":"2025-11-28T15:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.911915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.911976 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.911994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.912016 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:16 crc kubenswrapper[4805]: I1128 15:27:16.912034 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:16Z","lastTransitionTime":"2025-11-28T15:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.015301 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.015337 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.015348 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.015378 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.015388 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.118508 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.118550 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.118562 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.118577 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.118587 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.203708 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:17 crc kubenswrapper[4805]: E1128 15:27:17.204066 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.204991 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.205049 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:17 crc kubenswrapper[4805]: E1128 15:27:17.205226 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:17 crc kubenswrapper[4805]: E1128 15:27:17.205418 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.221476 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.221528 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.221545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.221572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.221590 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.324470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.324532 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.324554 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.324584 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.324607 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.427304 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.427343 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.427380 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.427400 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.427413 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.530243 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.530280 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.530291 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.530308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.530320 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.633658 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.633694 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.633703 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.633718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.633728 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.736103 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.736136 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.736144 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.736159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.736170 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.840104 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.840155 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.840167 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.840184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.840199 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.943175 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.943237 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.943251 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.943273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:17 crc kubenswrapper[4805]: I1128 15:27:17.943287 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:17Z","lastTransitionTime":"2025-11-28T15:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.045523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.045577 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.045591 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.045617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.045632 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.148535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.148597 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.148613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.148638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.148656 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.203502 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:18 crc kubenswrapper[4805]: E1128 15:27:18.203615 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.251485 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.251527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.251570 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.251596 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.251607 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.357717 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.357791 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.357806 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.357834 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.357852 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.407079 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:18 crc kubenswrapper[4805]: E1128 15:27:18.407300 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:27:18 crc kubenswrapper[4805]: E1128 15:27:18.407644 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs podName:18cf186f-76f6-47bc-8db9-c9b7be3aaf09 nodeName:}" failed. No retries permitted until 2025-11-28 15:27:50.407619358 +0000 UTC m=+97.457410669 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs") pod "network-metrics-daemon-fplc8" (UID: "18cf186f-76f6-47bc-8db9-c9b7be3aaf09") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.461262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.461340 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.461406 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.461443 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.461469 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.564532 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.564583 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.564594 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.564615 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.564629 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.667980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.668032 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.668043 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.668062 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.668074 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.771559 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.771596 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.771607 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.771625 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.771639 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.874488 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.874525 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.874534 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.874547 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.874558 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.977236 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.977290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.977299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.977315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:18 crc kubenswrapper[4805]: I1128 15:27:18.977325 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:18Z","lastTransitionTime":"2025-11-28T15:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.081076 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.081171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.081184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.081222 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.081235 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:19Z","lastTransitionTime":"2025-11-28T15:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.184218 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.184271 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.184283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.184299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.184310 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:19Z","lastTransitionTime":"2025-11-28T15:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.203062 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.203106 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.203172 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:19 crc kubenswrapper[4805]: E1128 15:27:19.203222 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:19 crc kubenswrapper[4805]: E1128 15:27:19.203332 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:19 crc kubenswrapper[4805]: E1128 15:27:19.203464 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.286503 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.286570 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.286581 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.286629 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.286645 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:19Z","lastTransitionTime":"2025-11-28T15:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.389590 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.389638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.389653 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.389670 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.389680 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:19Z","lastTransitionTime":"2025-11-28T15:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.492844 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.492885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.492897 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.492917 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.492931 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:19Z","lastTransitionTime":"2025-11-28T15:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.596031 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.596086 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.596099 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.596121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.596134 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:19Z","lastTransitionTime":"2025-11-28T15:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.664183 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/0.log" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.664227 4805 generic.go:334] "Generic (PLEG): container finished" podID="9d0031c5-0433-419e-9363-66eb48341a68" containerID="64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5" exitCode=1 Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.664257 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fv2dw" event={"ID":"9d0031c5-0433-419e-9363-66eb48341a68","Type":"ContainerDied","Data":"64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.664593 4805 scope.go:117] "RemoveContainer" containerID="64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.686447 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.698436 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.698755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.698767 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.698788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.698799 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:19Z","lastTransitionTime":"2025-11-28T15:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.702666 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.719073 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.730717 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.745768 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.757430 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.769393 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.782765 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.792687 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"030619ba-de4f-4db2-8138-4744768d40a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.800973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.801121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.801219 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.801303 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.801417 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:19Z","lastTransitionTime":"2025-11-28T15:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.804496 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.823469 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.835937 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.848044 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.870608 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.884351 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.894936 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.905058 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.910148 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.910186 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.910210 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.910225 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.910235 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:19Z","lastTransitionTime":"2025-11-28T15:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:19 crc kubenswrapper[4805]: I1128 15:27:19.916115 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:18Z\\\",\\\"message\\\":\\\"2025-11-28T15:26:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b\\\\n2025-11-28T15:26:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b to /host/opt/cni/bin/\\\\n2025-11-28T15:26:33Z [verbose] multus-daemon started\\\\n2025-11-28T15:26:33Z [verbose] Readiness Indicator file check\\\\n2025-11-28T15:27:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:19Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.013298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.013365 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.013375 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.013391 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.013401 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.115991 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.116040 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.116054 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.116070 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.116081 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.203378 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:20 crc kubenswrapper[4805]: E1128 15:27:20.203531 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.218417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.218456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.218466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.218482 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.218493 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.321412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.321459 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.321471 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.321488 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.321503 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.424477 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.424525 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.424535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.424553 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.424565 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.527004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.527043 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.527052 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.527067 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.527078 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.630630 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.630674 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.630686 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.630702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.630713 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.671100 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/0.log" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.671198 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fv2dw" event={"ID":"9d0031c5-0433-419e-9363-66eb48341a68","Type":"ContainerStarted","Data":"d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.696986 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.709176 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.719930 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.732824 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.732931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.732941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.732956 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.732967 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.733768 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.747778 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:18Z\\\",\\\"message\\\":\\\"2025-11-28T15:26:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b\\\\n2025-11-28T15:26:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b to /host/opt/cni/bin/\\\\n2025-11-28T15:26:33Z [verbose] multus-daemon started\\\\n2025-11-28T15:26:33Z [verbose] Readiness Indicator file check\\\\n2025-11-28T15:27:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.757431 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.768281 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.786892 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.798874 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.813721 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.827614 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.835738 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.835767 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.835776 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.835790 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.835800 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.840134 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.849569 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.860778 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.872568 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.883177 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"030619ba-de4f-4db2-8138-4744768d40a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.894745 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.915576 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:20Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.937858 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.937885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.937893 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.937907 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:20 crc kubenswrapper[4805]: I1128 15:27:20.937916 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:20Z","lastTransitionTime":"2025-11-28T15:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.040082 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.040118 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.040126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.040140 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.040149 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.143349 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.143947 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.144030 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.144171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.144265 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.203284 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.203315 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:21 crc kubenswrapper[4805]: E1128 15:27:21.203476 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.203542 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:21 crc kubenswrapper[4805]: E1128 15:27:21.203657 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:21 crc kubenswrapper[4805]: E1128 15:27:21.203773 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.247779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.247853 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.247877 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.247911 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.247933 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.350991 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.351297 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.351396 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.351482 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.351560 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.454294 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.454339 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.454350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.454386 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.454398 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.557813 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.557863 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.557874 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.557894 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.557907 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.660287 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.660340 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.660377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.660400 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.660418 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.762927 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.762969 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.762979 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.762997 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.763010 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.866480 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.866512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.866520 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.866557 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.866567 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.969299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.969347 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.969374 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.969389 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:21 crc kubenswrapper[4805]: I1128 15:27:21.969400 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:21Z","lastTransitionTime":"2025-11-28T15:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.071703 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.071742 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.071751 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.071765 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.071773 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.174034 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.174099 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.174109 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.174127 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.174136 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.203657 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:22 crc kubenswrapper[4805]: E1128 15:27:22.203801 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.276271 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.276338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.276356 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.276401 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.276419 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.379883 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.379951 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.379974 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.380010 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.380037 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.482875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.482963 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.482993 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.483029 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.483048 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.585925 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.585980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.585992 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.586016 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.586031 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.689424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.689487 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.689500 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.689527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.689538 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.792311 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.792375 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.792387 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.792427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.792440 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.894754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.894816 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.894829 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.894845 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.894857 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.997732 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.997769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.997780 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.997797 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:22 crc kubenswrapper[4805]: I1128 15:27:22.997809 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:22Z","lastTransitionTime":"2025-11-28T15:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.100223 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.100272 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.100286 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.100303 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.100314 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.202863 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.202932 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:23 crc kubenswrapper[4805]: E1128 15:27:23.202995 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.203175 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:23 crc kubenswrapper[4805]: E1128 15:27:23.203182 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.203314 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.203413 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.203442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.203467 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.203485 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: E1128 15:27:23.203387 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.215878 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.229160 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.243819 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:18Z\\\",\\\"message\\\":\\\"2025-11-28T15:26:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b\\\\n2025-11-28T15:26:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b to /host/opt/cni/bin/\\\\n2025-11-28T15:26:33Z [verbose] multus-daemon started\\\\n2025-11-28T15:26:33Z [verbose] Readiness Indicator file check\\\\n2025-11-28T15:27:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.248539 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.248574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.248585 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.248603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.248615 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.254406 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: E1128 15:27:23.262334 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.264716 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.265982 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.266025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.266033 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.266048 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.266056 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: E1128 15:27:23.277484 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.281858 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.281914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.281954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.281970 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.281980 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.283312 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: E1128 15:27:23.295942 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.296548 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.299348 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.299383 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.299391 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.299404 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.299413 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.310616 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: E1128 15:27:23.311812 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.314693 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.314718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.314725 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.314737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.314747 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.321729 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: E1128 15:27:23.324878 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: E1128 15:27:23.324978 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.326624 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.326646 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.326679 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.326693 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.326703 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.332203 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.343611 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.354071 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.364482 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.374693 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.386190 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.402099 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.412283 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"030619ba-de4f-4db2-8138-4744768d40a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.422479 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:23Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.429180 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.429219 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.429229 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.429245 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.429257 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.531705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.531740 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.531748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.531761 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.531770 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.634043 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.634105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.634122 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.634148 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.634166 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.736883 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.736923 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.736935 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.736951 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.736963 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.839752 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.839818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.839836 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.839863 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.839881 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.942151 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.942334 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.942348 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.942383 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:23 crc kubenswrapper[4805]: I1128 15:27:23.942395 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:23Z","lastTransitionTime":"2025-11-28T15:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.044730 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.044769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.044802 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.044821 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.044833 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.147768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.147857 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.147870 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.147888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.147899 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.203581 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:24 crc kubenswrapper[4805]: E1128 15:27:24.203685 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.249932 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.249984 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.249994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.250008 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.250017 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.352543 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.352614 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.352633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.352661 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.352679 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.455692 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.455755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.455771 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.455796 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.455826 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.558429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.558498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.558518 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.558545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.558566 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.661080 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.661132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.661144 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.661164 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.661178 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.763911 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.763958 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.763969 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.763986 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.763998 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.867056 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.867120 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.867174 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.867199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.867216 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.969804 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.969914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.969959 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.969982 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:24 crc kubenswrapper[4805]: I1128 15:27:24.969993 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:24Z","lastTransitionTime":"2025-11-28T15:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.072455 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.072513 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.072527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.072542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.072551 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:25Z","lastTransitionTime":"2025-11-28T15:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.175741 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.175778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.175788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.175804 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.175813 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:25Z","lastTransitionTime":"2025-11-28T15:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.203836 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.203911 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:25 crc kubenswrapper[4805]: E1128 15:27:25.203967 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.203990 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:25 crc kubenswrapper[4805]: E1128 15:27:25.204029 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:25 crc kubenswrapper[4805]: E1128 15:27:25.204135 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.278750 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.278823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.278845 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.278872 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.278896 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:25Z","lastTransitionTime":"2025-11-28T15:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.382076 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.382126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.382136 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.382152 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.382160 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:25Z","lastTransitionTime":"2025-11-28T15:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.484790 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.484816 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.484823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.484837 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.484846 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:25Z","lastTransitionTime":"2025-11-28T15:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.586739 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.586771 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.586779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.586793 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.586802 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:25Z","lastTransitionTime":"2025-11-28T15:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.690854 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.690898 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.690911 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.690928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.690944 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:25Z","lastTransitionTime":"2025-11-28T15:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.792869 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.792900 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.792910 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.792925 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.792937 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:25Z","lastTransitionTime":"2025-11-28T15:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.899006 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.899260 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.899273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.899291 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:25 crc kubenswrapper[4805]: I1128 15:27:25.899302 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:25Z","lastTransitionTime":"2025-11-28T15:27:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.000970 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.001199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.001213 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.001227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.001236 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.103620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.103668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.103680 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.103697 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.103709 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.203056 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:26 crc kubenswrapper[4805]: E1128 15:27:26.203191 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.206121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.206161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.206171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.206189 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.206199 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.308446 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.308492 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.308504 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.308523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.308537 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.411154 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.411220 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.411241 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.411272 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.411293 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.513249 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.513290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.513300 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.513313 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.513322 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.616696 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.616747 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.616763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.616785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.616801 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.720089 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.720148 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.720159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.720178 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.720194 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.823412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.823455 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.823468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.823487 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.823502 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.926694 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.926753 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.926783 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.926822 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:26 crc kubenswrapper[4805]: I1128 15:27:26.926844 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:26Z","lastTransitionTime":"2025-11-28T15:27:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.030521 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.030581 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.030596 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.030622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.030638 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.133420 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.133479 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.133495 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.133515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.133531 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.203147 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.203198 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.203153 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:27 crc kubenswrapper[4805]: E1128 15:27:27.203285 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:27 crc kubenswrapper[4805]: E1128 15:27:27.203555 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:27 crc kubenswrapper[4805]: E1128 15:27:27.203688 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.237311 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.237429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.237451 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.237476 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.237492 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.340441 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.340505 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.340522 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.340547 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.340595 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.443674 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.443768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.443805 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.443839 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.443863 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.547060 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.547098 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.547107 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.547122 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.547133 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.649674 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.649724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.649738 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.649759 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.649772 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.753053 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.753105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.753116 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.753135 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.753148 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.855636 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.855682 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.855698 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.855719 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.855733 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.958054 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.958099 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.958113 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.958133 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:27 crc kubenswrapper[4805]: I1128 15:27:27.958147 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:27Z","lastTransitionTime":"2025-11-28T15:27:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.060442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.060492 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.060507 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.060529 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.060544 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.163850 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.163914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.163932 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.163955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.163972 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.203425 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:28 crc kubenswrapper[4805]: E1128 15:27:28.203913 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.204107 4805 scope.go:117] "RemoveContainer" containerID="ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.272482 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.272742 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.272834 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.272866 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.272881 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.376397 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.376737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.376868 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.376955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.377044 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.480834 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.481192 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.481262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.481342 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.481443 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.584883 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.584924 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.584939 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.584964 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.584981 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.688070 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.688143 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.688165 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.688191 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.688211 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.701028 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/2.log" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.704121 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.790991 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.791030 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.791042 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.791062 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.791075 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.894279 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.894328 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.894341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.894378 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.894396 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.997106 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.997143 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.997152 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.997169 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:28 crc kubenswrapper[4805]: I1128 15:27:28.997178 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:28Z","lastTransitionTime":"2025-11-28T15:27:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.099786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.099832 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.099843 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.099862 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.099878 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:29Z","lastTransitionTime":"2025-11-28T15:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.202770 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.202812 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.202825 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.202845 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.203037 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:29Z","lastTransitionTime":"2025-11-28T15:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.202913 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.202978 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:29 crc kubenswrapper[4805]: E1128 15:27:29.203215 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.202913 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:29 crc kubenswrapper[4805]: E1128 15:27:29.203383 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:29 crc kubenswrapper[4805]: E1128 15:27:29.203399 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.305722 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.305765 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.305775 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.305790 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.305799 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:29Z","lastTransitionTime":"2025-11-28T15:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.408258 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.408290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.408298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.408314 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.408323 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:29Z","lastTransitionTime":"2025-11-28T15:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.511152 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.511226 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.511244 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.511275 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.511302 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:29Z","lastTransitionTime":"2025-11-28T15:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.614170 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.614343 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.614408 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.614439 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.614460 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:29Z","lastTransitionTime":"2025-11-28T15:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.708659 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.717285 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.717340 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.717386 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.717413 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.717430 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:29Z","lastTransitionTime":"2025-11-28T15:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.729655 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.750211 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.765455 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.775915 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.789638 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.803746 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.820073 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.820123 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.820136 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.820159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.820172 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:29Z","lastTransitionTime":"2025-11-28T15:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.827727 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.841350 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"030619ba-de4f-4db2-8138-4744768d40a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.856234 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.872314 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.887108 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.902081 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:18Z\\\",\\\"message\\\":\\\"2025-11-28T15:26:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b\\\\n2025-11-28T15:26:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b to /host/opt/cni/bin/\\\\n2025-11-28T15:26:33Z [verbose] multus-daemon started\\\\n2025-11-28T15:26:33Z [verbose] Readiness Indicator file check\\\\n2025-11-28T15:27:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.914633 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.923658 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.923688 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.923696 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.923710 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.923719 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:29Z","lastTransitionTime":"2025-11-28T15:27:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.926831 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.947071 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.957224 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.971529 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:29 crc kubenswrapper[4805]: I1128 15:27:29.983163 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.026875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.026918 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.026932 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.026948 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.026960 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.129804 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.129868 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.129890 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.129921 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.129945 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.203720 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:30 crc kubenswrapper[4805]: E1128 15:27:30.203870 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.232130 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.232168 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.232176 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.232188 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.232197 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.345123 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.345193 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.345212 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.345239 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.345301 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.448168 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.448249 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.448262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.448283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.448295 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.550583 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.550607 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.550615 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.550631 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.550641 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.654020 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.654090 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.654114 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.654145 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.654169 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.713778 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/3.log" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.714578 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/2.log" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.718336 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" exitCode=1 Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.718394 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.718430 4805 scope.go:117] "RemoveContainer" containerID="ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.719085 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:27:30 crc kubenswrapper[4805]: E1128 15:27:30.719274 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.734618 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.749928 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.756384 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.756432 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.756445 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.756462 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.756473 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.763493 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.775398 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.785524 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.795751 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"030619ba-de4f-4db2-8138-4744768d40a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.807789 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.824338 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba3e4b57205b1564e78fc6ce6b71b1416f9a6acd1ab81abbd983ca1624aaf3ae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:00Z\\\",\\\"message\\\":\\\"rk-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 15:27:00.002317 6439 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1128 15:27:00.002327 6439 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1128 15:27:00.002341 6439 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1128 15:27:00.002330 6439 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_respo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:29Z\\\",\\\"message\\\":\\\"org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1128 15:27:29.326195 6804 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:27:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.834184 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.843998 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.858128 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.858179 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.858196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.858242 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.858259 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.864378 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.875539 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.890914 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.901824 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.915261 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:18Z\\\",\\\"message\\\":\\\"2025-11-28T15:26:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b\\\\n2025-11-28T15:26:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b to /host/opt/cni/bin/\\\\n2025-11-28T15:26:33Z [verbose] multus-daemon started\\\\n2025-11-28T15:26:33Z [verbose] Readiness Indicator file check\\\\n2025-11-28T15:27:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.925557 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.935182 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.946548 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:30Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.960842 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.960880 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.960888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.960912 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:30 crc kubenswrapper[4805]: I1128 15:27:30.960923 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:30Z","lastTransitionTime":"2025-11-28T15:27:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.071256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.071341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.071371 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.071393 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.071407 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:31Z","lastTransitionTime":"2025-11-28T15:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.175124 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.175168 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.175182 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.175199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.175217 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:31Z","lastTransitionTime":"2025-11-28T15:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.203857 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.203940 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:31 crc kubenswrapper[4805]: E1128 15:27:31.204041 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.204103 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:31 crc kubenswrapper[4805]: E1128 15:27:31.204246 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:31 crc kubenswrapper[4805]: E1128 15:27:31.204285 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.277633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.277671 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.277681 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.277696 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.277707 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:31Z","lastTransitionTime":"2025-11-28T15:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.381127 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.381298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.381417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.381457 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.381482 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:31Z","lastTransitionTime":"2025-11-28T15:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.484571 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.484623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.484634 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.484651 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.484661 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:31Z","lastTransitionTime":"2025-11-28T15:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.587142 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.587431 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.587575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.587774 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.587936 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:31Z","lastTransitionTime":"2025-11-28T15:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.691567 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.691621 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.691638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.691664 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.691685 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:31Z","lastTransitionTime":"2025-11-28T15:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.723306 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:27:31 crc kubenswrapper[4805]: E1128 15:27:31.723833 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.743231 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"030619ba-de4f-4db2-8138-4744768d40a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.766313 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.791864 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:29Z\\\",\\\"message\\\":\\\"org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1128 15:27:29.326195 6804 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:27:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.795713 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.795877 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.795926 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.795960 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.795989 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:31Z","lastTransitionTime":"2025-11-28T15:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.825906 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.847706 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.865597 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.884972 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.898729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.898819 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.898861 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.898895 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.898917 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:31Z","lastTransitionTime":"2025-11-28T15:27:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.902826 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:18Z\\\",\\\"message\\\":\\\"2025-11-28T15:26:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b\\\\n2025-11-28T15:26:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b to /host/opt/cni/bin/\\\\n2025-11-28T15:26:33Z [verbose] multus-daemon started\\\\n2025-11-28T15:26:33Z [verbose] Readiness Indicator file check\\\\n2025-11-28T15:27:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.918420 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.936144 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.951128 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.963597 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.980412 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:31 crc kubenswrapper[4805]: I1128 15:27:31.994043 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:31Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.001864 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.001919 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.001928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.001942 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.001974 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.009945 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.025583 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.042103 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.058303 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:32Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.105068 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.105112 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.105125 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.105143 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.105156 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.203718 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:32 crc kubenswrapper[4805]: E1128 15:27:32.203883 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.207853 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.207892 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.207902 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.207915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.207925 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.312020 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.312090 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.312113 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.312136 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.312154 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.414722 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.414773 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.414785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.414807 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.414818 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.518148 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.518635 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.518807 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.518963 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.519121 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.622091 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.622428 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.622540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.622623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.622718 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.730886 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.731273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.731298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.731321 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.731335 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.736114 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/3.log" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.834893 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.835175 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.835245 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.835310 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.835406 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.937553 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.937799 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.937867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.937938 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:32 crc kubenswrapper[4805]: I1128 15:27:32.938004 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:32Z","lastTransitionTime":"2025-11-28T15:27:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.042002 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.042064 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.042080 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.042105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.042121 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.144460 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.144510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.144523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.144544 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.144557 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.203887 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.203939 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.204233 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:33 crc kubenswrapper[4805]: E1128 15:27:33.204315 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:33 crc kubenswrapper[4805]: E1128 15:27:33.204461 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:33 crc kubenswrapper[4805]: E1128 15:27:33.204650 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.214843 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.221476 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.234811 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.247670 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.247748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.247760 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.247780 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.247794 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.248703 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.259185 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.272419 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.286809 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.305463 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:29Z\\\",\\\"message\\\":\\\"org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1128 15:27:29.326195 6804 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:27:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.318198 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"030619ba-de4f-4db2-8138-4744768d40a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.329968 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.340265 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.351095 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.351259 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.351289 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.351307 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.351324 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.351334 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.363136 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:18Z\\\",\\\"message\\\":\\\"2025-11-28T15:26:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b\\\\n2025-11-28T15:26:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b to /host/opt/cni/bin/\\\\n2025-11-28T15:26:33Z [verbose] multus-daemon started\\\\n2025-11-28T15:26:33Z [verbose] Readiness Indicator file check\\\\n2025-11-28T15:27:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.371989 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.383014 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.398884 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.409069 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.421600 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.434246 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.454282 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.454547 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.454688 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.454879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.455071 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.496859 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.497083 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.497167 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.497276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.497386 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: E1128 15:27:33.510328 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.514051 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.514088 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.514098 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.514114 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.514125 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: E1128 15:27:33.524423 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.527669 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.527700 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.527709 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.527724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.527735 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: E1128 15:27:33.542904 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.546815 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.546846 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.546854 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.546867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.546877 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: E1128 15:27:33.560914 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.564961 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.564999 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.565007 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.565022 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.565033 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: E1128 15:27:33.578572 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:33Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:33 crc kubenswrapper[4805]: E1128 15:27:33.578711 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.580416 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.580447 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.580456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.580470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.580480 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.682594 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.683061 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.683264 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.683544 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.683751 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.787393 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.787805 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.788077 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.788312 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.788574 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.891821 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.891885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.891907 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.891935 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.891958 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.996147 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.996678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.996847 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.997004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:33 crc kubenswrapper[4805]: I1128 15:27:33.997149 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:33Z","lastTransitionTime":"2025-11-28T15:27:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.099965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.100403 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.100655 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.100848 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.101170 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:34Z","lastTransitionTime":"2025-11-28T15:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.203025 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:34 crc kubenswrapper[4805]: E1128 15:27:34.203174 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.203789 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.203817 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.203827 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.203842 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.203854 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:34Z","lastTransitionTime":"2025-11-28T15:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.306698 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.306752 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.306769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.306800 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.306824 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:34Z","lastTransitionTime":"2025-11-28T15:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.410268 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.410317 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.410328 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.410346 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.410376 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:34Z","lastTransitionTime":"2025-11-28T15:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.514175 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.514222 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.514234 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.514251 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.514265 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:34Z","lastTransitionTime":"2025-11-28T15:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.617199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.617239 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.617248 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.617264 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.617276 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:34Z","lastTransitionTime":"2025-11-28T15:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.720502 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.720562 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.720579 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.720604 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.720621 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:34Z","lastTransitionTime":"2025-11-28T15:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.823795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.823878 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.823900 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.823929 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.823953 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:34Z","lastTransitionTime":"2025-11-28T15:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.927944 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.927988 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.928000 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.928021 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:34 crc kubenswrapper[4805]: I1128 15:27:34.928034 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:34Z","lastTransitionTime":"2025-11-28T15:27:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.030469 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.030501 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.030510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.030527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.030539 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.092014 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.092102 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:39.092085687 +0000 UTC m=+146.141876998 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.132612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.132646 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.132655 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.132669 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.132677 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.193168 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.193468 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.193498 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.193515 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193334 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193626 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193624 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193667 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:28:39.193654628 +0000 UTC m=+146.243445939 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193672 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193719 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 15:28:39.193702129 +0000 UTC m=+146.243493460 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193587 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193755 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 15:28:39.193746381 +0000 UTC m=+146.243537712 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193582 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193774 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193785 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.193809 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 15:28:39.193800672 +0000 UTC m=+146.243591993 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.203355 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.203416 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.203515 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.203608 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.203706 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:35 crc kubenswrapper[4805]: E1128 15:27:35.203948 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.235177 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.235221 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.235234 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.235253 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.235267 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.338173 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.338201 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.338208 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.338221 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.338229 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.440811 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.440855 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.440867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.440882 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.440894 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.543379 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.543421 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.543432 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.543448 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.543460 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.646305 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.646341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.646349 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.646377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.646386 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.748668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.748763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.748798 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.748837 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.748861 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.851536 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.851622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.851645 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.851678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.851702 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.954026 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.954075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.954088 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.954106 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:35 crc kubenswrapper[4805]: I1128 15:27:35.954119 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:35Z","lastTransitionTime":"2025-11-28T15:27:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.057105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.057159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.057172 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.057190 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.057202 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.159979 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.160022 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.160035 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.160053 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.160067 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.203269 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:36 crc kubenswrapper[4805]: E1128 15:27:36.203470 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.262599 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.262646 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.262657 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.262675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.262686 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.366047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.366121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.366143 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.366172 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.366191 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.468514 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.468558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.468569 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.468588 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.468602 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.571658 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.571768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.571791 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.571820 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.571841 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.674439 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.674597 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.674627 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.674654 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.674672 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.777753 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.777818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.777835 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.777863 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.777881 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.882033 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.882075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.882086 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.882102 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.882114 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.985457 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.985505 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.985515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.985535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:36 crc kubenswrapper[4805]: I1128 15:27:36.985548 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:36Z","lastTransitionTime":"2025-11-28T15:27:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.088756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.088794 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.088805 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.088822 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.088837 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:37Z","lastTransitionTime":"2025-11-28T15:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.192144 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.192221 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.192245 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.192277 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.192302 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:37Z","lastTransitionTime":"2025-11-28T15:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.203869 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.203906 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.203987 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:37 crc kubenswrapper[4805]: E1128 15:27:37.204258 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:37 crc kubenswrapper[4805]: E1128 15:27:37.204485 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:37 crc kubenswrapper[4805]: E1128 15:27:37.204553 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.295346 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.295418 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.295433 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.295457 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.295473 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:37Z","lastTransitionTime":"2025-11-28T15:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.401081 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.401145 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.401158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.401179 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.401191 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:37Z","lastTransitionTime":"2025-11-28T15:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.503386 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.503444 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.503456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.503474 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.503487 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:37Z","lastTransitionTime":"2025-11-28T15:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.606063 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.606101 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.606110 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.606125 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.606136 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:37Z","lastTransitionTime":"2025-11-28T15:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.709193 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.709729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.709761 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.709795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.709819 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:37Z","lastTransitionTime":"2025-11-28T15:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.812236 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.812287 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.812298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.812317 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.812329 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:37Z","lastTransitionTime":"2025-11-28T15:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.915848 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.915930 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.915949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.915973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:37 crc kubenswrapper[4805]: I1128 15:27:37.915991 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:37Z","lastTransitionTime":"2025-11-28T15:27:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.019651 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.019727 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.019769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.019802 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.019824 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.123506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.123592 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.123614 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.123643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.123665 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.203765 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:38 crc kubenswrapper[4805]: E1128 15:27:38.203943 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.226206 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.226269 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.226282 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.226302 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.226315 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.328818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.328861 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.328872 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.328886 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.328899 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.430803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.430848 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.430858 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.430877 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.430888 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.533908 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.533959 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.533969 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.533989 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.534001 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.636130 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.636178 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.636187 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.636203 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.636213 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.740618 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.740655 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.740665 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.740681 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.740693 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.843427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.843479 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.843495 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.843515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.843528 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.945215 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.945261 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.945279 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.945298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:38 crc kubenswrapper[4805]: I1128 15:27:38.945323 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:38Z","lastTransitionTime":"2025-11-28T15:27:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.047023 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.047076 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.047094 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.047116 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.047132 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.149757 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.149790 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.149799 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.149813 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.149823 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.203838 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.203934 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:39 crc kubenswrapper[4805]: E1128 15:27:39.203976 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.203840 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:39 crc kubenswrapper[4805]: E1128 15:27:39.204093 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:39 crc kubenswrapper[4805]: E1128 15:27:39.204178 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.252376 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.252650 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.252744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.252842 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.252942 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.355266 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.355306 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.355319 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.355337 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.355348 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.458153 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.458196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.458211 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.458229 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.458243 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.561826 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.561866 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.561880 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.561899 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.561913 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.664086 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.664699 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.664745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.664766 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.664779 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.767524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.767575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.767589 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.767612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.767627 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.870557 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.870602 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.870613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.870633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.870645 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.973443 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.973523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.973540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.973567 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:39 crc kubenswrapper[4805]: I1128 15:27:39.973586 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:39Z","lastTransitionTime":"2025-11-28T15:27:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.076957 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.077008 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.077022 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.077050 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.077064 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:40Z","lastTransitionTime":"2025-11-28T15:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.180765 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.180811 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.180828 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.180852 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.180869 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:40Z","lastTransitionTime":"2025-11-28T15:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.203601 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:40 crc kubenswrapper[4805]: E1128 15:27:40.203758 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.284607 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.284678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.284698 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.284729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.284751 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:40Z","lastTransitionTime":"2025-11-28T15:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.386544 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.386582 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.386593 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.386610 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.386619 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:40Z","lastTransitionTime":"2025-11-28T15:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.488757 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.488994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.489065 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.489140 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.489210 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:40Z","lastTransitionTime":"2025-11-28T15:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.595453 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.595536 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.595556 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.595678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.596263 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:40Z","lastTransitionTime":"2025-11-28T15:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.699483 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.699571 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.699588 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.699609 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.699623 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:40Z","lastTransitionTime":"2025-11-28T15:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.802701 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.802938 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.803004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.803100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.803193 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:40Z","lastTransitionTime":"2025-11-28T15:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.906099 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.906167 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.906184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.906201 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:40 crc kubenswrapper[4805]: I1128 15:27:40.906211 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:40Z","lastTransitionTime":"2025-11-28T15:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.008445 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.008481 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.008492 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.008510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.008521 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.111678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.111718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.111728 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.111742 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.111753 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.203860 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.203930 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:41 crc kubenswrapper[4805]: E1128 15:27:41.203999 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:41 crc kubenswrapper[4805]: E1128 15:27:41.204074 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.204185 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:41 crc kubenswrapper[4805]: E1128 15:27:41.204428 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.214711 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.214803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.214824 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.215205 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.215227 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.318700 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.318746 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.318756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.318772 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.318785 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.420769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.420803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.420812 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.420825 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.420835 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.523505 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.523550 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.523562 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.523578 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.523590 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.626744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.626784 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.626794 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.626810 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.626819 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.729096 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.729142 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.729154 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.729169 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.729180 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.831227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.831271 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.831283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.831300 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.831311 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.933894 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.933930 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.933941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.933956 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:41 crc kubenswrapper[4805]: I1128 15:27:41.933967 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:41Z","lastTransitionTime":"2025-11-28T15:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.037581 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.037626 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.037637 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.037654 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.037669 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.139957 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.140020 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.140033 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.140051 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.140063 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.203162 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:42 crc kubenswrapper[4805]: E1128 15:27:42.203357 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.242631 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.242691 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.242714 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.242744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.242766 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.345015 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.345083 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.345100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.345126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.345146 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.447905 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.447966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.447984 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.448007 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.448025 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.550235 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.550530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.550624 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.550754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.550845 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.672521 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.672607 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.672633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.672666 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.672689 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.774842 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.774879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.774892 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.774908 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.774920 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.878101 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.878176 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.878197 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.878230 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.878252 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.982001 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.982111 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.982132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.982159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:42 crc kubenswrapper[4805]: I1128 15:27:42.982177 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:42Z","lastTransitionTime":"2025-11-28T15:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.084459 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.084548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.084579 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.084649 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.084673 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.188872 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.189205 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.189300 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.189429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.189552 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.203074 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.203191 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.203069 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:43 crc kubenswrapper[4805]: E1128 15:27:43.203471 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:43 crc kubenswrapper[4805]: E1128 15:27:43.203552 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:43 crc kubenswrapper[4805]: E1128 15:27:43.203613 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.216430 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7688879f-c48b-4b0b-9aa5-2bde152cf34b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d95c91e030bf6514e9308ca7e6aa389a50cfdb1225814f0945e8f11471b750f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a644ca713d38c92c80c20f5f8a3010932e7e7471bd4b312b2e02b53798e525fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2pg6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-64s9v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.243983 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e9362ab-5afc-4a82-b2b2-fa8ef48b37c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1743f474f44ad927d7cdcc3af33300efc7a075c7c81576ba31acbe168618f8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf839b7fd336032529c321e0399dd0327f07e8481f726e93ad4003827638073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://850e1dc719825a7ed2f33f6f01f8051be4e68e0dd7a75aa0d9f243442764ca2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a08e8058bcbc00489a9fb5cf6d0c0402e16c39decea1cc6109747fc6c0934883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd69987408be464cd7e8bc7fd8cc0a4b9820a844aadbfc8eab298e1ab7396e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be0a2f20956c46c76f3cb575e13b4f7e36238cea23a190983f3f1574e83a1944\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca1a18eed658d47d866d2c9c4b66f2fe3f92d045e2c572fa1aaa8cec334990fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69efcefa881eb8ce5697f30c8b48e70031d99ac0e1634b9187a33d996708e8f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.259314 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8f42d3-5dd5-49c9-8645-382075c8e289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 15:26:25.468908 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 15:26:25.470350 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-665292694/tls.crt::/tmp/serving-cert-665292694/tls.key\\\\\\\"\\\\nI1128 15:26:31.189306 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 15:26:31.191532 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 15:26:31.191548 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 15:26:31.191566 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 15:26:31.191571 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 15:26:31.196256 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 15:26:31.196273 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196278 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 15:26:31.196283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 15:26:31.196286 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 15:26:31.196289 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 15:26:31.196291 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1128 15:26:31.196340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1128 15:26:31.198603 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.271889 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.286330 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://822796ebfec1c12d9079cec4a5084acf51aa827e05af1a1d14fcfa871b7bb71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.292691 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.292737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.292753 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.292774 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.292785 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.300633 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fv2dw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0031c5-0433-419e-9363-66eb48341a68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:18Z\\\",\\\"message\\\":\\\"2025-11-28T15:26:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b\\\\n2025-11-28T15:26:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_80a34786-d82f-4471-96a6-f5255ffd566b to /host/opt/cni/bin/\\\\n2025-11-28T15:26:33Z [verbose] multus-daemon started\\\\n2025-11-28T15:26:33Z [verbose] Readiness Indicator file check\\\\n2025-11-28T15:27:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ztm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fv2dw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.311992 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jmhd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d26ccf-e51d-4b2d-a17f-766fa06e71ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e5af753aada69250387f10b4149a27e3607303f5003898db9ef39e01d7f643e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nh2vb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jmhd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.326576 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af9ca51b-34ad-419d-a201-eb91904b799b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32e75caf9f69cf0804002e3086989594318d5fdde36eadaa3c45c8b70f76cc0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://704b377b01f0fb2c99cc8de04d9df397e31ebe4bb82b5ee46044e91b52310ce9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018723768bba2bc61b2b6a898867aed8c2aca867479fdb5cd4df2bcd75a62fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.339943 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73da70d8-a7dc-4fca-9e65-9c0d0d815966\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://844ad4a75ef9b31ad113894ba673325a4e76025b5adda72a0c0ad09c7063af87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ftdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9pzmp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.354637 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e746e8b8-a375-4650-98fc-f6e3ba35a50d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa0e903f6576b2148284ee2da294d9a1e93e2b183aa9b22ccda5d74981825a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a11a3e83324690808aeebc541fcba52f1d659a52bcd278fd56966a603f97a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d27af4e1b45d037898b48276a25f6af863957174cfb717077d8462881e196b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://044a95b385d7b09e02ef165d2ba8da71aba91f9e46c033916e42c085c4731044\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e06db4e85a4883740b72d219ed4ed1f82e109294e2236be97f20fff403a4ab0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://271b6707e6aa29a26ca108285ba0d48717a7bff059ef916fb2b97b07f2768874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77aceb147ba190bc51594552c82a16bcfbc96a71e6af19f7b745c1f983e87ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mf4k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wdzqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.364463 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df807f5-4a87-4549-8d3b-4ce1a99300af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649e8e6f083503baab3d95c91d1f7e86d02a1e6521b91f703fef65d6d7820803\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c10fa94785110aee1f90c069bbc2aa7ac5ab39aa6382988c291972ddea33deb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c10fa94785110aee1f90c069bbc2aa7ac5ab39aa6382988c291972ddea33deb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.381137 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.392882 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.394705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.394768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.394793 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.394824 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.394843 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.404075 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82299169d6091c07e18d9efdacdaa369c5abce5e6eb12e1c1b6231d498f394b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.416953 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tjqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f70dd0f8-a9a1-4694-827a-6c27deed4080\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7429256f9589520f50d50ff498280b299d4802f4d4a4f69cc387058b0887c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mkgd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tjqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.428192 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fplc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fplc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.442232 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"030619ba-de4f-4db2-8138-4744768d40a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a19f86b8234447058be04b634215cf7ef8e99734684798cb926b3a4707855f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e04a15599fcc1246308149e6688ae91dc1703671a39cce65ee7cfc71cdcb4ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://028e90e45da79376fed66c0b22d94edd6fe3ac484bd4971a64991e9027ae9d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346daa40bfed74a0be0a50626f50190bae4eb228cf725706eda1d0780aa85e8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.455854 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://222f551024d926467ae9589075f6cc0c438afd2e71513723e1baff16b6066184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://673b6ae34f8a010f0b7598a170507afa843cd480e2cc4f727fa1364f103319af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.484474 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"412627f3-6ef3-401d-b7ae-a839d70a46b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:26:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T15:27:29Z\\\",\\\"message\\\":\\\"org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1128 15:27:29.326195 6804 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:29Z is after 2025-0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T15:27:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T15:26:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T15:26:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T15:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpjph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:26:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6v4kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.497701 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.497750 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.497765 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.497786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.497800 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.600140 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.600616 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.600855 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.601073 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.601280 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.613649 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.613749 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.613783 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.613818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.613838 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: E1128 15:27:43.636152 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.641307 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.641349 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.641388 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.641406 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.641415 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: E1128 15:27:43.658605 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.662828 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.662881 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.662896 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.662918 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.662933 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: E1128 15:27:43.675012 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.679230 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.679274 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.679286 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.679306 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.679320 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: E1128 15:27:43.692709 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.698424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.698462 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.698471 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.698486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.698497 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: E1128 15:27:43.711682 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T15:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b34b13c7-e3bf-4272-b34c-e52b7c30974a\\\",\\\"systemUUID\\\":\\\"164ad39d-af70-4732-b323-1a0ca23607cd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T15:27:43Z is after 2025-08-24T17:21:41Z" Nov 28 15:27:43 crc kubenswrapper[4805]: E1128 15:27:43.712353 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.714048 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.714082 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.714093 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.714111 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.714123 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.816152 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.816199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.816224 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.816250 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.816270 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.919755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.919836 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.919860 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.919897 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:43 crc kubenswrapper[4805]: I1128 15:27:43.919921 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:43Z","lastTransitionTime":"2025-11-28T15:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.022649 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.022685 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.022694 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.022707 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.022725 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.125783 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.126137 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.126436 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.126674 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.126894 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.203231 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:44 crc kubenswrapper[4805]: E1128 15:27:44.203392 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.205172 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:27:44 crc kubenswrapper[4805]: E1128 15:27:44.205704 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.230197 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.230254 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.230265 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.230281 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.230290 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.333764 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.333863 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.333887 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.333922 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.333949 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.438865 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.438960 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.438983 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.439016 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.439038 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.541393 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.541483 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.541498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.541517 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.541531 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.643585 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.643662 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.643686 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.643716 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.643737 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.746807 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.747172 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.747298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.747456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.747577 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.851333 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.852117 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.852157 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.852187 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.852249 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.955065 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.955131 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.955149 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.955176 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:44 crc kubenswrapper[4805]: I1128 15:27:44.955191 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:44Z","lastTransitionTime":"2025-11-28T15:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.058411 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.058468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.058484 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.058511 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.058529 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.162334 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.162506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.162541 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.162574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.162595 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.203683 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.203873 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.203693 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:45 crc kubenswrapper[4805]: E1128 15:27:45.204023 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:45 crc kubenswrapper[4805]: E1128 15:27:45.204161 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:45 crc kubenswrapper[4805]: E1128 15:27:45.204221 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.265535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.265600 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.265617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.265643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.265659 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.368534 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.368600 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.368617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.368642 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.368659 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.472267 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.472326 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.472338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.472360 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.472396 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.575588 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.575640 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.575656 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.575676 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.575690 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.678334 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.678410 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.678423 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.678442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.678459 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.780121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.780189 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.780202 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.780218 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.780230 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.883034 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.883094 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.883110 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.883132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.883151 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.986407 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.986680 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.986747 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.986816 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:45 crc kubenswrapper[4805]: I1128 15:27:45.986876 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:45Z","lastTransitionTime":"2025-11-28T15:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.089188 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.089246 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.089260 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.089281 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.089294 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:46Z","lastTransitionTime":"2025-11-28T15:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.193646 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.193707 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.193725 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.193752 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.193772 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:46Z","lastTransitionTime":"2025-11-28T15:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.203638 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:46 crc kubenswrapper[4805]: E1128 15:27:46.204174 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.297690 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.297726 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.297739 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.297757 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.297770 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:46Z","lastTransitionTime":"2025-11-28T15:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.400915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.400969 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.400984 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.401004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.401016 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:46Z","lastTransitionTime":"2025-11-28T15:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.503754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.503829 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.503849 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.503881 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.503903 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:46Z","lastTransitionTime":"2025-11-28T15:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.608057 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.608149 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.608162 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.608234 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.608258 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:46Z","lastTransitionTime":"2025-11-28T15:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.710705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.710745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.710756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.710774 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.710790 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:46Z","lastTransitionTime":"2025-11-28T15:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.813953 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.814050 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.814073 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.814108 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.814139 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:46Z","lastTransitionTime":"2025-11-28T15:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.916268 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.916319 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.916330 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.916346 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:46 crc kubenswrapper[4805]: I1128 15:27:46.916384 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:46Z","lastTransitionTime":"2025-11-28T15:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.019437 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.019497 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.019511 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.019537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.019550 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.121682 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.121795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.121816 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.121879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.121898 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.203783 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.203819 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:47 crc kubenswrapper[4805]: E1128 15:27:47.204013 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:47 crc kubenswrapper[4805]: E1128 15:27:47.204231 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.204592 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:47 crc kubenswrapper[4805]: E1128 15:27:47.204821 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.224887 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.224966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.224986 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.225016 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.225037 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.329265 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.329320 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.329335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.329382 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.329402 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.432129 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.432198 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.432216 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.432244 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.432261 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.537120 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.537176 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.537187 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.537206 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.537217 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.640510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.640553 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.640569 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.640588 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.640605 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.743622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.743898 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.744041 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.744165 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.744283 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.846440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.846493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.846509 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.846530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.846548 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.949895 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.949942 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.949953 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.949971 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:47 crc kubenswrapper[4805]: I1128 15:27:47.949986 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:47Z","lastTransitionTime":"2025-11-28T15:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.053190 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.053327 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.053387 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.053429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.053515 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.156873 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.156923 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.156939 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.156962 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.156974 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.203507 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:48 crc kubenswrapper[4805]: E1128 15:27:48.203753 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.260251 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.260290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.260303 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.260321 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.260336 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.363648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.364196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.364312 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.364453 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.364557 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.468150 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.468213 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.468232 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.468256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.468277 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.570980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.571032 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.571048 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.571071 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.571087 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.673980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.674055 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.674079 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.674105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.674121 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.777430 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.777485 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.777497 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.777517 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.777530 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.880673 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.880718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.880728 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.880748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.880757 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.984305 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.984386 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.984399 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.984420 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:48 crc kubenswrapper[4805]: I1128 15:27:48.984434 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:48Z","lastTransitionTime":"2025-11-28T15:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.087418 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.087483 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.087498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.087521 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.087541 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:49Z","lastTransitionTime":"2025-11-28T15:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.190718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.190796 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.190820 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.190859 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.190884 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:49Z","lastTransitionTime":"2025-11-28T15:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.203405 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.203504 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.203798 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:49 crc kubenswrapper[4805]: E1128 15:27:49.203992 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:49 crc kubenswrapper[4805]: E1128 15:27:49.204096 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:49 crc kubenswrapper[4805]: E1128 15:27:49.204424 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.299724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.299796 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.299822 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.299863 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.299885 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:49Z","lastTransitionTime":"2025-11-28T15:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.402640 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.402689 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.402707 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.402731 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.402748 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:49Z","lastTransitionTime":"2025-11-28T15:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.506045 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.506440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.506620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.506778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.506915 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:49Z","lastTransitionTime":"2025-11-28T15:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.609190 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.609229 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.609237 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.609252 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.609260 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:49Z","lastTransitionTime":"2025-11-28T15:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.712212 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.712268 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.712284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.712310 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.712329 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:49Z","lastTransitionTime":"2025-11-28T15:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.816056 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.816637 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.816901 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.817089 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.817224 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:49Z","lastTransitionTime":"2025-11-28T15:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.920694 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.920747 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.920763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.920788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:49 crc kubenswrapper[4805]: I1128 15:27:49.920808 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:49Z","lastTransitionTime":"2025-11-28T15:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.024574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.024612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.024622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.024642 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.024654 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.128994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.130107 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.130159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.130191 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.130213 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.203074 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:50 crc kubenswrapper[4805]: E1128 15:27:50.203304 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.233504 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.233594 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.233612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.233639 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.233658 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.335726 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.335780 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.335800 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.335824 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.335842 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.438920 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.439003 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.439025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.439049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.439068 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.476344 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:50 crc kubenswrapper[4805]: E1128 15:27:50.476534 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:27:50 crc kubenswrapper[4805]: E1128 15:27:50.477157 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs podName:18cf186f-76f6-47bc-8db9-c9b7be3aaf09 nodeName:}" failed. No retries permitted until 2025-11-28 15:28:54.477136505 +0000 UTC m=+161.526927826 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs") pod "network-metrics-daemon-fplc8" (UID: "18cf186f-76f6-47bc-8db9-c9b7be3aaf09") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.541290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.541350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.541375 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.541392 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.541403 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.644825 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.644879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.644893 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.644912 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.644925 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.748503 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.748553 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.748574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.748599 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.748618 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.851383 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.851431 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.851440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.851454 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.851463 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.953349 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.953401 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.953413 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.953427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:50 crc kubenswrapper[4805]: I1128 15:27:50.953440 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:50Z","lastTransitionTime":"2025-11-28T15:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.055805 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.055850 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.055867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.055887 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.055898 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.158654 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.158695 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.158707 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.158729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.158741 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.203601 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.203672 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.203737 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:51 crc kubenswrapper[4805]: E1128 15:27:51.203748 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:51 crc kubenswrapper[4805]: E1128 15:27:51.203828 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:51 crc kubenswrapper[4805]: E1128 15:27:51.204010 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.261511 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.261547 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.261555 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.261566 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.261574 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.370110 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.370174 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.370192 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.370216 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.370235 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.474047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.474122 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.474145 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.474173 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.474195 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.576996 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.577049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.577066 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.577095 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.577114 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.680424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.680493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.680514 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.680540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.680557 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.783916 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.783975 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.783993 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.784019 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.784037 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.886742 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.886792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.886804 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.886823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.886835 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.990424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.990888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.991043 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.991174 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:51 crc kubenswrapper[4805]: I1128 15:27:51.991326 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:51Z","lastTransitionTime":"2025-11-28T15:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.093734 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.093790 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.093813 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.093842 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.093864 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:52Z","lastTransitionTime":"2025-11-28T15:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.196415 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.196478 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.196497 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.196522 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.196540 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:52Z","lastTransitionTime":"2025-11-28T15:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.203829 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:52 crc kubenswrapper[4805]: E1128 15:27:52.204007 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.299070 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.299500 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.299679 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.299865 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.300063 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:52Z","lastTransitionTime":"2025-11-28T15:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.403468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.403533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.403559 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.403591 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.403614 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:52Z","lastTransitionTime":"2025-11-28T15:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.505944 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.506011 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.506033 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.506064 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.506084 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:52Z","lastTransitionTime":"2025-11-28T15:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.608779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.608868 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.608893 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.608932 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.608959 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:52Z","lastTransitionTime":"2025-11-28T15:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.723068 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.723118 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.723131 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.723160 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.723176 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:52Z","lastTransitionTime":"2025-11-28T15:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.825519 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.825603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.825618 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.825638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.825651 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:52Z","lastTransitionTime":"2025-11-28T15:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.927949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.927986 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.927998 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.928015 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:52 crc kubenswrapper[4805]: I1128 15:27:52.928027 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:52Z","lastTransitionTime":"2025-11-28T15:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.030737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.030810 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.030836 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.030866 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.030890 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.133966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.134011 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.134024 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.134043 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.134061 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.203343 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.203422 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:53 crc kubenswrapper[4805]: E1128 15:27:53.203526 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:53 crc kubenswrapper[4805]: E1128 15:27:53.203613 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.203789 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:53 crc kubenswrapper[4805]: E1128 15:27:53.203907 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.237646 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.237709 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.237721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.237737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.237747 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.272953 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-wdzqc" podStartSLOduration=81.272926192 podStartE2EDuration="1m21.272926192s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.245669343 +0000 UTC m=+100.295460674" watchObservedRunningTime="2025-11-28 15:27:53.272926192 +0000 UTC m=+100.322717543" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.308998 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=82.308977242 podStartE2EDuration="1m22.308977242s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.275156655 +0000 UTC m=+100.324947976" watchObservedRunningTime="2025-11-28 15:27:53.308977242 +0000 UTC m=+100.358768573" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.339096 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podStartSLOduration=82.339077161 podStartE2EDuration="1m22.339077161s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.309199239 +0000 UTC m=+100.358990610" watchObservedRunningTime="2025-11-28 15:27:53.339077161 +0000 UTC m=+100.388868472" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.340201 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.340259 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.340270 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.340285 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.340309 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.360072 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-9tjqf" podStartSLOduration=82.360048281 podStartE2EDuration="1m22.360048281s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.359844405 +0000 UTC m=+100.409635716" watchObservedRunningTime="2025-11-28 15:27:53.360048281 +0000 UTC m=+100.409839612" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.380267 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=20.380249738 podStartE2EDuration="20.380249738s" podCreationTimestamp="2025-11-28 15:27:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.379892018 +0000 UTC m=+100.429683339" watchObservedRunningTime="2025-11-28 15:27:53.380249738 +0000 UTC m=+100.430041049" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.442907 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.442949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.442965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.443002 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.443014 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.449754 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.449735913 podStartE2EDuration="45.449735913s" podCreationTimestamp="2025-11-28 15:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.4489001 +0000 UTC m=+100.498691411" watchObservedRunningTime="2025-11-28 15:27:53.449735913 +0000 UTC m=+100.499527224" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.499017 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-fv2dw" podStartSLOduration=82.49899843 podStartE2EDuration="1m22.49899843s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.498419154 +0000 UTC m=+100.548210475" watchObservedRunningTime="2025-11-28 15:27:53.49899843 +0000 UTC m=+100.548789741" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.522322 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-jmhd2" podStartSLOduration=82.522305776 podStartE2EDuration="1m22.522305776s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.509798479 +0000 UTC m=+100.559589810" watchObservedRunningTime="2025-11-28 15:27:53.522305776 +0000 UTC m=+100.572097087" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.522841 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-64s9v" podStartSLOduration=81.522838101 podStartE2EDuration="1m21.522838101s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.522621455 +0000 UTC m=+100.572412816" watchObservedRunningTime="2025-11-28 15:27:53.522838101 +0000 UTC m=+100.572629412" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.545526 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.545567 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.545578 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.545596 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.545607 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.547211 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=81.547197118 podStartE2EDuration="1m21.547197118s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.544880231 +0000 UTC m=+100.594671562" watchObservedRunningTime="2025-11-28 15:27:53.547197118 +0000 UTC m=+100.596988429" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.565041 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=82.565022836 podStartE2EDuration="1m22.565022836s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:53.563962467 +0000 UTC m=+100.613753778" watchObservedRunningTime="2025-11-28 15:27:53.565022836 +0000 UTC m=+100.614814147" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.647399 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.647433 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.647441 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.647454 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.647463 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.749755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.749793 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.749925 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.749965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.750005 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.853072 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.853105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.853112 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.853124 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.853133 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.955949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.956006 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.956023 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.956045 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.956061 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.996703 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.996810 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.996829 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.996891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 15:27:53 crc kubenswrapper[4805]: I1128 15:27:53.996912 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T15:27:53Z","lastTransitionTime":"2025-11-28T15:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.065140 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj"] Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.065538 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.068770 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.068798 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.068804 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.069956 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.203832 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:54 crc kubenswrapper[4805]: E1128 15:27:54.204086 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.240319 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d840615b-e5ca-4330-be10-777185fd9eeb-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.240389 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d840615b-e5ca-4330-be10-777185fd9eeb-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.240422 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d840615b-e5ca-4330-be10-777185fd9eeb-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.240533 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d840615b-e5ca-4330-be10-777185fd9eeb-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.240686 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d840615b-e5ca-4330-be10-777185fd9eeb-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.341948 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d840615b-e5ca-4330-be10-777185fd9eeb-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.342002 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d840615b-e5ca-4330-be10-777185fd9eeb-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.342039 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d840615b-e5ca-4330-be10-777185fd9eeb-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.342060 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d840615b-e5ca-4330-be10-777185fd9eeb-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.342103 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d840615b-e5ca-4330-be10-777185fd9eeb-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.342597 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d840615b-e5ca-4330-be10-777185fd9eeb-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.342757 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d840615b-e5ca-4330-be10-777185fd9eeb-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.343094 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d840615b-e5ca-4330-be10-777185fd9eeb-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.349220 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d840615b-e5ca-4330-be10-777185fd9eeb-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.358297 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d840615b-e5ca-4330-be10-777185fd9eeb-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5ddzj\" (UID: \"d840615b-e5ca-4330-be10-777185fd9eeb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.378629 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.819138 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" event={"ID":"d840615b-e5ca-4330-be10-777185fd9eeb","Type":"ContainerStarted","Data":"102fc576924f4d7a0060ded7b5cb22b03c088a5d79f0bcae5e1a33700ef60999"} Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.819550 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" event={"ID":"d840615b-e5ca-4330-be10-777185fd9eeb","Type":"ContainerStarted","Data":"f560db7f29016992723fcf7c359f86d44c12dd8cd1e3ad2ab473b9e16d9fa960"} Nov 28 15:27:54 crc kubenswrapper[4805]: I1128 15:27:54.839448 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5ddzj" podStartSLOduration=83.839423473 podStartE2EDuration="1m23.839423473s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:27:54.839422463 +0000 UTC m=+101.889213794" watchObservedRunningTime="2025-11-28 15:27:54.839423473 +0000 UTC m=+101.889214824" Nov 28 15:27:55 crc kubenswrapper[4805]: I1128 15:27:55.203456 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:55 crc kubenswrapper[4805]: I1128 15:27:55.203534 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:55 crc kubenswrapper[4805]: I1128 15:27:55.203690 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:55 crc kubenswrapper[4805]: E1128 15:27:55.203673 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:55 crc kubenswrapper[4805]: E1128 15:27:55.203839 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:55 crc kubenswrapper[4805]: E1128 15:27:55.204078 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:56 crc kubenswrapper[4805]: I1128 15:27:56.203792 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:56 crc kubenswrapper[4805]: E1128 15:27:56.204200 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:57 crc kubenswrapper[4805]: I1128 15:27:57.203536 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:57 crc kubenswrapper[4805]: I1128 15:27:57.203918 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:57 crc kubenswrapper[4805]: I1128 15:27:57.204060 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:57 crc kubenswrapper[4805]: E1128 15:27:57.204383 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:27:57 crc kubenswrapper[4805]: E1128 15:27:57.204523 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:57 crc kubenswrapper[4805]: E1128 15:27:57.205644 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:58 crc kubenswrapper[4805]: I1128 15:27:58.203533 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:27:58 crc kubenswrapper[4805]: E1128 15:27:58.203873 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:27:58 crc kubenswrapper[4805]: I1128 15:27:58.204084 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:27:58 crc kubenswrapper[4805]: E1128 15:27:58.204229 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6v4kb_openshift-ovn-kubernetes(412627f3-6ef3-401d-b7ae-a839d70a46b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" Nov 28 15:27:59 crc kubenswrapper[4805]: I1128 15:27:59.203500 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:27:59 crc kubenswrapper[4805]: I1128 15:27:59.203500 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:27:59 crc kubenswrapper[4805]: E1128 15:27:59.204490 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:27:59 crc kubenswrapper[4805]: I1128 15:27:59.203689 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:27:59 crc kubenswrapper[4805]: E1128 15:27:59.204696 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:27:59 crc kubenswrapper[4805]: E1128 15:27:59.204857 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:00 crc kubenswrapper[4805]: I1128 15:28:00.203495 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:00 crc kubenswrapper[4805]: E1128 15:28:00.203701 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:01 crc kubenswrapper[4805]: I1128 15:28:01.203890 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:01 crc kubenswrapper[4805]: I1128 15:28:01.203963 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:01 crc kubenswrapper[4805]: I1128 15:28:01.203969 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:01 crc kubenswrapper[4805]: E1128 15:28:01.204979 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:01 crc kubenswrapper[4805]: E1128 15:28:01.205107 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:01 crc kubenswrapper[4805]: E1128 15:28:01.205048 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:02 crc kubenswrapper[4805]: I1128 15:28:02.203607 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:02 crc kubenswrapper[4805]: E1128 15:28:02.203880 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:03 crc kubenswrapper[4805]: I1128 15:28:03.204621 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:03 crc kubenswrapper[4805]: I1128 15:28:03.204746 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:03 crc kubenswrapper[4805]: E1128 15:28:03.206684 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:03 crc kubenswrapper[4805]: I1128 15:28:03.206756 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:03 crc kubenswrapper[4805]: E1128 15:28:03.206828 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:03 crc kubenswrapper[4805]: E1128 15:28:03.206929 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:04 crc kubenswrapper[4805]: I1128 15:28:04.203533 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:04 crc kubenswrapper[4805]: E1128 15:28:04.203770 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:05 crc kubenswrapper[4805]: I1128 15:28:05.203837 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:05 crc kubenswrapper[4805]: E1128 15:28:05.203970 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:05 crc kubenswrapper[4805]: I1128 15:28:05.204004 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:05 crc kubenswrapper[4805]: E1128 15:28:05.204137 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:05 crc kubenswrapper[4805]: I1128 15:28:05.204024 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:05 crc kubenswrapper[4805]: E1128 15:28:05.204462 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:05 crc kubenswrapper[4805]: I1128 15:28:05.869729 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/1.log" Nov 28 15:28:05 crc kubenswrapper[4805]: I1128 15:28:05.870462 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/0.log" Nov 28 15:28:05 crc kubenswrapper[4805]: I1128 15:28:05.870542 4805 generic.go:334] "Generic (PLEG): container finished" podID="9d0031c5-0433-419e-9363-66eb48341a68" containerID="d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55" exitCode=1 Nov 28 15:28:05 crc kubenswrapper[4805]: I1128 15:28:05.870589 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fv2dw" event={"ID":"9d0031c5-0433-419e-9363-66eb48341a68","Type":"ContainerDied","Data":"d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55"} Nov 28 15:28:05 crc kubenswrapper[4805]: I1128 15:28:05.870647 4805 scope.go:117] "RemoveContainer" containerID="64c1c524f0f02711bc13fb96e96b7d648c602ee6767e10d19f6489455192e9b5" Nov 28 15:28:05 crc kubenswrapper[4805]: I1128 15:28:05.871010 4805 scope.go:117] "RemoveContainer" containerID="d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55" Nov 28 15:28:05 crc kubenswrapper[4805]: E1128 15:28:05.872127 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-fv2dw_openshift-multus(9d0031c5-0433-419e-9363-66eb48341a68)\"" pod="openshift-multus/multus-fv2dw" podUID="9d0031c5-0433-419e-9363-66eb48341a68" Nov 28 15:28:06 crc kubenswrapper[4805]: I1128 15:28:06.203007 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:06 crc kubenswrapper[4805]: E1128 15:28:06.203201 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:06 crc kubenswrapper[4805]: I1128 15:28:06.876933 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/1.log" Nov 28 15:28:07 crc kubenswrapper[4805]: I1128 15:28:07.203684 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:07 crc kubenswrapper[4805]: I1128 15:28:07.203761 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:07 crc kubenswrapper[4805]: I1128 15:28:07.203811 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:07 crc kubenswrapper[4805]: E1128 15:28:07.203991 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:07 crc kubenswrapper[4805]: E1128 15:28:07.204322 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:07 crc kubenswrapper[4805]: E1128 15:28:07.204690 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:08 crc kubenswrapper[4805]: I1128 15:28:08.203106 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:08 crc kubenswrapper[4805]: E1128 15:28:08.203306 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:09 crc kubenswrapper[4805]: I1128 15:28:09.203233 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:09 crc kubenswrapper[4805]: I1128 15:28:09.203494 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:09 crc kubenswrapper[4805]: E1128 15:28:09.203758 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:09 crc kubenswrapper[4805]: E1128 15:28:09.204007 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:09 crc kubenswrapper[4805]: I1128 15:28:09.204282 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:09 crc kubenswrapper[4805]: E1128 15:28:09.204465 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:10 crc kubenswrapper[4805]: I1128 15:28:10.203717 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:10 crc kubenswrapper[4805]: E1128 15:28:10.203854 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:11 crc kubenswrapper[4805]: I1128 15:28:11.203480 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:11 crc kubenswrapper[4805]: I1128 15:28:11.203529 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:11 crc kubenswrapper[4805]: I1128 15:28:11.203531 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:11 crc kubenswrapper[4805]: E1128 15:28:11.203682 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:11 crc kubenswrapper[4805]: E1128 15:28:11.203768 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:11 crc kubenswrapper[4805]: E1128 15:28:11.203871 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:12 crc kubenswrapper[4805]: I1128 15:28:12.203412 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:12 crc kubenswrapper[4805]: E1128 15:28:12.203613 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:13 crc kubenswrapper[4805]: E1128 15:28:13.168662 4805 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 28 15:28:13 crc kubenswrapper[4805]: I1128 15:28:13.203124 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:13 crc kubenswrapper[4805]: E1128 15:28:13.205756 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:13 crc kubenswrapper[4805]: I1128 15:28:13.205870 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:13 crc kubenswrapper[4805]: I1128 15:28:13.205897 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:13 crc kubenswrapper[4805]: E1128 15:28:13.206858 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:13 crc kubenswrapper[4805]: E1128 15:28:13.206869 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:13 crc kubenswrapper[4805]: I1128 15:28:13.206958 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:28:13 crc kubenswrapper[4805]: E1128 15:28:13.306130 4805 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 15:28:13 crc kubenswrapper[4805]: I1128 15:28:13.906414 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/3.log" Nov 28 15:28:13 crc kubenswrapper[4805]: I1128 15:28:13.909203 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerStarted","Data":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} Nov 28 15:28:13 crc kubenswrapper[4805]: I1128 15:28:13.909807 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:28:13 crc kubenswrapper[4805]: I1128 15:28:13.944033 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podStartSLOduration=101.944012841 podStartE2EDuration="1m41.944012841s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:13.943780804 +0000 UTC m=+120.993572155" watchObservedRunningTime="2025-11-28 15:28:13.944012841 +0000 UTC m=+120.993804172" Nov 28 15:28:14 crc kubenswrapper[4805]: I1128 15:28:14.090723 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fplc8"] Nov 28 15:28:14 crc kubenswrapper[4805]: I1128 15:28:14.090848 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:14 crc kubenswrapper[4805]: E1128 15:28:14.090948 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:15 crc kubenswrapper[4805]: I1128 15:28:15.203866 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:15 crc kubenswrapper[4805]: I1128 15:28:15.203953 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:15 crc kubenswrapper[4805]: E1128 15:28:15.204244 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:15 crc kubenswrapper[4805]: I1128 15:28:15.203955 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:15 crc kubenswrapper[4805]: E1128 15:28:15.204353 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:15 crc kubenswrapper[4805]: I1128 15:28:15.203953 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:15 crc kubenswrapper[4805]: E1128 15:28:15.204483 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:15 crc kubenswrapper[4805]: E1128 15:28:15.204547 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:17 crc kubenswrapper[4805]: I1128 15:28:17.203812 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:17 crc kubenswrapper[4805]: I1128 15:28:17.203850 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:17 crc kubenswrapper[4805]: I1128 15:28:17.203885 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:17 crc kubenswrapper[4805]: I1128 15:28:17.203832 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:17 crc kubenswrapper[4805]: E1128 15:28:17.204048 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:17 crc kubenswrapper[4805]: E1128 15:28:17.204217 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:17 crc kubenswrapper[4805]: E1128 15:28:17.204338 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:17 crc kubenswrapper[4805]: E1128 15:28:17.204524 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:18 crc kubenswrapper[4805]: E1128 15:28:18.307468 4805 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 15:28:19 crc kubenswrapper[4805]: I1128 15:28:19.204142 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:19 crc kubenswrapper[4805]: I1128 15:28:19.205019 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:19 crc kubenswrapper[4805]: I1128 15:28:19.205043 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:19 crc kubenswrapper[4805]: E1128 15:28:19.205211 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:19 crc kubenswrapper[4805]: I1128 15:28:19.205556 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:19 crc kubenswrapper[4805]: E1128 15:28:19.206012 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:19 crc kubenswrapper[4805]: E1128 15:28:19.206178 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:19 crc kubenswrapper[4805]: E1128 15:28:19.206263 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:19 crc kubenswrapper[4805]: I1128 15:28:19.206632 4805 scope.go:117] "RemoveContainer" containerID="d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55" Nov 28 15:28:19 crc kubenswrapper[4805]: I1128 15:28:19.934602 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/1.log" Nov 28 15:28:19 crc kubenswrapper[4805]: I1128 15:28:19.934678 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fv2dw" event={"ID":"9d0031c5-0433-419e-9363-66eb48341a68","Type":"ContainerStarted","Data":"7b9dc44d838a23520473b186116f6a70691671146dfb822f52de1795381db630"} Nov 28 15:28:21 crc kubenswrapper[4805]: I1128 15:28:21.203647 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:21 crc kubenswrapper[4805]: I1128 15:28:21.203648 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:21 crc kubenswrapper[4805]: E1128 15:28:21.204241 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:21 crc kubenswrapper[4805]: I1128 15:28:21.203814 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:21 crc kubenswrapper[4805]: I1128 15:28:21.203777 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:21 crc kubenswrapper[4805]: E1128 15:28:21.204415 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:21 crc kubenswrapper[4805]: E1128 15:28:21.204587 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:21 crc kubenswrapper[4805]: E1128 15:28:21.204691 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:23 crc kubenswrapper[4805]: I1128 15:28:23.203265 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:23 crc kubenswrapper[4805]: I1128 15:28:23.203413 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:23 crc kubenswrapper[4805]: I1128 15:28:23.203504 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:23 crc kubenswrapper[4805]: E1128 15:28:23.204538 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 15:28:23 crc kubenswrapper[4805]: I1128 15:28:23.204560 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:23 crc kubenswrapper[4805]: E1128 15:28:23.204711 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 15:28:23 crc kubenswrapper[4805]: E1128 15:28:23.204764 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 15:28:23 crc kubenswrapper[4805]: E1128 15:28:23.204886 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fplc8" podUID="18cf186f-76f6-47bc-8db9-c9b7be3aaf09" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.514424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.559768 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.560102 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wltgm"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.560425 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.561044 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.563015 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.564027 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.564449 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.564814 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v7h7c"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.565572 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vjvw8"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.565653 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.566320 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.566404 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.566800 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.566980 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q6czs"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.567165 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.567873 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-scnxw"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.568067 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.568559 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.568641 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jgwwr"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.569086 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.569173 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.581900 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.585956 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.586342 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.586967 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.588215 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.588304 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.588372 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.588970 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.589173 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.589180 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.589738 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.589812 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-7bprj"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.589983 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.590215 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.590285 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.590535 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.590576 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.593553 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.594545 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.595332 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.596443 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.596473 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.596628 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.597270 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.615690 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.616919 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.616943 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.617093 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.617269 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.617274 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.617097 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.617613 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.617773 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.618029 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.618236 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.618239 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.618796 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.618881 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.619000 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-2b2hq"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.619540 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2b2hq" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.624591 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.624621 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.624691 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.624774 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.624622 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.624934 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.624968 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.625033 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.625892 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.626410 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.629236 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.629522 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.629754 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.629898 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.629754 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630032 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630124 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630144 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630223 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630329 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630342 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630457 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630527 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630565 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630593 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630697 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630718 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630719 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630806 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630854 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630875 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.630947 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.631218 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.631230 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.631323 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.631436 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.631562 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.631562 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.634566 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.634933 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vlp4p"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.635185 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.636179 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.636454 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.636643 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.637435 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pmls5"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.638177 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.638671 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pzmqp"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.639208 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.642443 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.642946 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.652607 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.653249 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.653493 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.654112 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.656689 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.656823 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.658605 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.681503 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.681711 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.683088 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5vvq"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.683790 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.684119 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.684835 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.684871 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.685042 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.685207 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.685633 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.685777 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.685903 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.686407 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.686800 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.686843 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/828da568-03e9-42d6-883b-7bbb1d429825-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.686872 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.686911 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43ead517-eb06-4b99-848f-7c49af380b6b-serving-cert\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.686933 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjcrj\" (UniqueName: \"kubernetes.io/projected/2946e237-1d94-4eaa-ba11-148add5a3843-kube-api-access-hjcrj\") pod \"openshift-config-operator-7777fb866f-wltgm\" (UID: \"2946e237-1d94-4eaa-ba11-148add5a3843\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.686954 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvrhs\" (UniqueName: \"kubernetes.io/projected/aebbc881-3380-4df9-b082-f6e0d16b8f8e-kube-api-access-kvrhs\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.686976 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/69d90cc8-5227-4516-9537-5d3b83f92dbd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.686997 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687021 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7305c6-5d47-4c06-aaa9-b4308c578ae6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ktchr\" (UID: \"2b7305c6-5d47-4c06-aaa9-b4308c578ae6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687042 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687061 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687173 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687202 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687248 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687287 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687062 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687307 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687327 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-audit\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687348 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f8xj\" (UniqueName: \"kubernetes.io/projected/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-kube-api-access-2f8xj\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687410 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2946e237-1d94-4eaa-ba11-148add5a3843-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wltgm\" (UID: \"2946e237-1d94-4eaa-ba11-148add5a3843\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687428 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687443 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687457 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/43ead517-eb06-4b99-848f-7c49af380b6b-node-pullsecrets\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687460 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687471 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-oauth-config\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687486 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687504 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwwjw\" (UniqueName: \"kubernetes.io/projected/97ea3cfc-6d2a-417f-a397-34a5760d0a9e-kube-api-access-fwwjw\") pod \"downloads-7954f5f757-2b2hq\" (UID: \"97ea3cfc-6d2a-417f-a397-34a5760d0a9e\") " pod="openshift-console/downloads-7954f5f757-2b2hq" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687520 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687549 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-config\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687562 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-trusted-ca-bundle\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687577 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/828da568-03e9-42d6-883b-7bbb1d429825-serving-cert\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687603 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687624 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-oauth-serving-cert\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687644 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b7305c6-5d47-4c06-aaa9-b4308c578ae6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ktchr\" (UID: \"2b7305c6-5d47-4c06-aaa9-b4308c578ae6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687661 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-dir\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687676 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e5cb325-16ec-41f9-8336-dcb96a7160cb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fw8v8\" (UID: \"7e5cb325-16ec-41f9-8336-dcb96a7160cb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687730 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-policies\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687754 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687801 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgtlb\" (UniqueName: \"kubernetes.io/projected/828da568-03e9-42d6-883b-7bbb1d429825-kube-api-access-qgtlb\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687821 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-serving-cert\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687836 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-audit-policies\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687858 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9e9a23de-9ba3-4581-aa91-dfb253372643-images\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687874 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6ncs\" (UniqueName: \"kubernetes.io/projected/9e9a23de-9ba3-4581-aa91-dfb253372643-kube-api-access-t6ncs\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687935 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-config\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.687992 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688001 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-etcd-serving-ca\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688038 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9e9a23de-9ba3-4581-aa91-dfb253372643-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688065 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nnmv\" (UniqueName: \"kubernetes.io/projected/43ead517-eb06-4b99-848f-7c49af380b6b-kube-api-access-4nnmv\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688090 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-config\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688179 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-encryption-config\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688236 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbvjv\" (UniqueName: \"kubernetes.io/projected/69d90cc8-5227-4516-9537-5d3b83f92dbd-kube-api-access-nbvjv\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688302 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828da568-03e9-42d6-883b-7bbb1d429825-config\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688327 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-service-ca\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688378 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjzxd\" (UniqueName: \"kubernetes.io/projected/69b46aab-afdf-4551-b0b4-b8d4f8f97831-kube-api-access-vjzxd\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688427 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/828da568-03e9-42d6-883b-7bbb1d429825-service-ca-bundle\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688465 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/69d90cc8-5227-4516-9537-5d3b83f92dbd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688492 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2946e237-1d94-4eaa-ba11-148add5a3843-serving-cert\") pod \"openshift-config-operator-7777fb866f-wltgm\" (UID: \"2946e237-1d94-4eaa-ba11-148add5a3843\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688552 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85065f7f-4262-49e8-8f55-4df7edc2b8a7-serving-cert\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688576 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-image-import-ca\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688611 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/69d90cc8-5227-4516-9537-5d3b83f92dbd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688635 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/43ead517-eb06-4b99-848f-7c49af380b6b-etcd-client\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688656 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688688 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzzck\" (UniqueName: \"kubernetes.io/projected/2b7305c6-5d47-4c06-aaa9-b4308c578ae6-kube-api-access-xzzck\") pod \"openshift-apiserver-operator-796bbdcf4f-ktchr\" (UID: \"2b7305c6-5d47-4c06-aaa9-b4308c578ae6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688730 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688757 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-serving-cert\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688810 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-etcd-client\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688837 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688858 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-audit-dir\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688886 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-client-ca\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688934 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/43ead517-eb06-4b99-848f-7c49af380b6b-audit-dir\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688962 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.688985 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnrsh\" (UniqueName: \"kubernetes.io/projected/85065f7f-4262-49e8-8f55-4df7edc2b8a7-kube-api-access-vnrsh\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.690892 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.690996 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9a23de-9ba3-4581-aa91-dfb253372643-config\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.691046 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/43ead517-eb06-4b99-848f-7c49af380b6b-encryption-config\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.691072 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84kzg\" (UniqueName: \"kubernetes.io/projected/7e5cb325-16ec-41f9-8336-dcb96a7160cb-kube-api-access-84kzg\") pod \"cluster-samples-operator-665b6dd947-fw8v8\" (UID: \"7e5cb325-16ec-41f9-8336-dcb96a7160cb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.691273 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.691478 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.691711 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.691998 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.692129 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.692185 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.693445 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.697610 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.698241 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.698301 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.699419 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.702707 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lfq9l"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.703222 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tfb5c"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.703376 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.703763 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.703861 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.705735 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.707616 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.708664 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.710233 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.710802 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.711738 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.712853 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.713889 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.714350 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dzb6r"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.714790 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.714854 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.715004 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.715312 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.717447 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.718071 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.719101 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.720064 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jq5cl"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.720128 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.720965 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.724161 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.724225 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.725079 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.725778 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.725834 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.726753 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.729111 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.734431 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.738213 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v7h7c"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.740163 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.744791 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pmls5"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.747118 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7bprj"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.748975 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.757076 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pzmqp"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.760861 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.764729 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vlp4p"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.764926 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.767456 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.769401 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.770619 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q6czs"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.772192 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.774131 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tfb5c"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.775443 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-scnxw"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.777697 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-bpmbs"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.778652 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.779486 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t4l7k"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.780267 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.781235 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2b2hq"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.782662 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vjvw8"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.782961 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.784525 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jgwwr"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.786117 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-trwpw"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.786579 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.787605 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-6j74g"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.787979 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6j74g" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.789589 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.790291 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.791604 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.791917 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/828da568-03e9-42d6-883b-7bbb1d429825-service-ca-bundle\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.791940 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/69d90cc8-5227-4516-9537-5d3b83f92dbd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.791957 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2946e237-1d94-4eaa-ba11-148add5a3843-serving-cert\") pod \"openshift-config-operator-7777fb866f-wltgm\" (UID: \"2946e237-1d94-4eaa-ba11-148add5a3843\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.791975 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e6d68711-24e1-46d1-95bb-31974c24b299-default-certificate\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.791992 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85065f7f-4262-49e8-8f55-4df7edc2b8a7-serving-cert\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792006 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-client-ca\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792022 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3158f6f2-98f3-4eb0-97d7-83c68dfeb743-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6dnsm\" (UID: \"3158f6f2-98f3-4eb0-97d7-83c68dfeb743\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792041 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/69d90cc8-5227-4516-9537-5d3b83f92dbd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792054 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/43ead517-eb06-4b99-848f-7c49af380b6b-etcd-client\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792069 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-image-import-ca\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792084 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlgwk\" (UniqueName: \"kubernetes.io/projected/e3be465c-283a-40aa-b004-5049a363ce29-kube-api-access-nlgwk\") pod \"dns-operator-744455d44c-pmls5\" (UID: \"e3be465c-283a-40aa-b004-5049a363ce29\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792099 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2dff72ce-7157-47d7-b48a-cf7e4c779d90-trusted-ca\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792115 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792130 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzzck\" (UniqueName: \"kubernetes.io/projected/2b7305c6-5d47-4c06-aaa9-b4308c578ae6-kube-api-access-xzzck\") pod \"openshift-apiserver-operator-796bbdcf4f-ktchr\" (UID: \"2b7305c6-5d47-4c06-aaa9-b4308c578ae6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792144 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2dff72ce-7157-47d7-b48a-cf7e4c779d90-metrics-tls\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792168 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792183 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-serving-cert\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792198 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-client-ca\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792219 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/43ead517-eb06-4b99-848f-7c49af380b6b-audit-dir\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792235 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-etcd-client\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792251 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792265 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-audit-dir\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792279 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792293 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnrsh\" (UniqueName: \"kubernetes.io/projected/85065f7f-4262-49e8-8f55-4df7edc2b8a7-kube-api-access-vnrsh\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792313 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-config\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792333 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9430f70-1c4a-47af-813f-76079af84e5e-serving-cert\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792368 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4zld\" (UniqueName: \"kubernetes.io/projected/e44b12c7-ec46-4a33-b9ac-87668a6465b5-kube-api-access-x4zld\") pod \"olm-operator-6b444d44fb-rzkmk\" (UID: \"e44b12c7-ec46-4a33-b9ac-87668a6465b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792389 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9a23de-9ba3-4581-aa91-dfb253372643-config\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792408 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/43ead517-eb06-4b99-848f-7c49af380b6b-encryption-config\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792429 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84kzg\" (UniqueName: \"kubernetes.io/projected/7e5cb325-16ec-41f9-8336-dcb96a7160cb-kube-api-access-84kzg\") pod \"cluster-samples-operator-665b6dd947-fw8v8\" (UID: \"7e5cb325-16ec-41f9-8336-dcb96a7160cb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792447 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792463 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/828da568-03e9-42d6-883b-7bbb1d429825-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792479 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792504 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvrhs\" (UniqueName: \"kubernetes.io/projected/aebbc881-3380-4df9-b082-f6e0d16b8f8e-kube-api-access-kvrhs\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792522 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/69d90cc8-5227-4516-9537-5d3b83f92dbd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792539 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43ead517-eb06-4b99-848f-7c49af380b6b-serving-cert\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792556 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjcrj\" (UniqueName: \"kubernetes.io/projected/2946e237-1d94-4eaa-ba11-148add5a3843-kube-api-access-hjcrj\") pod \"openshift-config-operator-7777fb866f-wltgm\" (UID: \"2946e237-1d94-4eaa-ba11-148add5a3843\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792573 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3be465c-283a-40aa-b004-5049a363ce29-metrics-tls\") pod \"dns-operator-744455d44c-pmls5\" (UID: \"e3be465c-283a-40aa-b004-5049a363ce29\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792611 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792627 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7305c6-5d47-4c06-aaa9-b4308c578ae6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ktchr\" (UID: \"2b7305c6-5d47-4c06-aaa9-b4308c578ae6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792642 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e44b12c7-ec46-4a33-b9ac-87668a6465b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-rzkmk\" (UID: \"e44b12c7-ec46-4a33-b9ac-87668a6465b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792658 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f8xj\" (UniqueName: \"kubernetes.io/projected/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-kube-api-access-2f8xj\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792674 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792691 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792706 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-audit\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792722 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e44b12c7-ec46-4a33-b9ac-87668a6465b5-srv-cert\") pod \"olm-operator-6b444d44fb-rzkmk\" (UID: \"e44b12c7-ec46-4a33-b9ac-87668a6465b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792738 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5999\" (UniqueName: \"kubernetes.io/projected/2dff72ce-7157-47d7-b48a-cf7e4c779d90-kube-api-access-m5999\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792756 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2946e237-1d94-4eaa-ba11-148add5a3843-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wltgm\" (UID: \"2946e237-1d94-4eaa-ba11-148add5a3843\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792772 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.792922 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/43ead517-eb06-4b99-848f-7c49af380b6b-audit-dir\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793074 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-audit-dir\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793645 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793668 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwwjw\" (UniqueName: \"kubernetes.io/projected/97ea3cfc-6d2a-417f-a397-34a5760d0a9e-kube-api-access-fwwjw\") pod \"downloads-7954f5f757-2b2hq\" (UID: \"97ea3cfc-6d2a-417f-a397-34a5760d0a9e\") " pod="openshift-console/downloads-7954f5f757-2b2hq" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793684 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/43ead517-eb06-4b99-848f-7c49af380b6b-node-pullsecrets\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793709 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-oauth-config\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793727 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793754 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793773 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-config\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793791 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-trusted-ca-bundle\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793807 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793821 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/828da568-03e9-42d6-883b-7bbb1d429825-serving-cert\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793837 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e6d68711-24e1-46d1-95bb-31974c24b299-stats-auth\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793852 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr2r8\" (UniqueName: \"kubernetes.io/projected/2c35f71e-388f-48ce-8e84-9d177a72978b-kube-api-access-wr2r8\") pod \"package-server-manager-789f6589d5-8zrtm\" (UID: \"2c35f71e-388f-48ce-8e84-9d177a72978b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793861 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-client-ca\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793876 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-oauth-serving-cert\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793892 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j79qp\" (UniqueName: \"kubernetes.io/projected/64014133-dfa9-4052-83a3-c88736be91dd-kube-api-access-j79qp\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793908 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6d68711-24e1-46d1-95bb-31974c24b299-service-ca-bundle\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793923 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-942lj\" (UniqueName: \"kubernetes.io/projected/e6d68711-24e1-46d1-95bb-31974c24b299-kube-api-access-942lj\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793945 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3158f6f2-98f3-4eb0-97d7-83c68dfeb743-config\") pod \"kube-controller-manager-operator-78b949d7b-6dnsm\" (UID: \"3158f6f2-98f3-4eb0-97d7-83c68dfeb743\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793960 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a59a18-ba76-4b57-bf78-d200d411f8b2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4zwjj\" (UID: \"53a59a18-ba76-4b57-bf78-d200d411f8b2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793976 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b7305c6-5d47-4c06-aaa9-b4308c578ae6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ktchr\" (UID: \"2b7305c6-5d47-4c06-aaa9-b4308c578ae6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793992 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-dir\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794007 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e5cb325-16ec-41f9-8336-dcb96a7160cb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fw8v8\" (UID: \"7e5cb325-16ec-41f9-8336-dcb96a7160cb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794023 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794040 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-policies\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794055 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgtlb\" (UniqueName: \"kubernetes.io/projected/828da568-03e9-42d6-883b-7bbb1d429825-kube-api-access-qgtlb\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794071 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-serving-cert\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794086 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-audit-policies\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794102 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/64014133-dfa9-4052-83a3-c88736be91dd-signing-key\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794116 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-config\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794132 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9e9a23de-9ba3-4581-aa91-dfb253372643-images\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794148 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6ncs\" (UniqueName: \"kubernetes.io/projected/9e9a23de-9ba3-4581-aa91-dfb253372643-kube-api-access-t6ncs\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794165 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-etcd-serving-ca\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794185 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c35f71e-388f-48ce-8e84-9d177a72978b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8zrtm\" (UID: \"2c35f71e-388f-48ce-8e84-9d177a72978b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794200 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2dff72ce-7157-47d7-b48a-cf7e4c779d90-bound-sa-token\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794216 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-config\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794234 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9e9a23de-9ba3-4581-aa91-dfb253372643-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794249 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nnmv\" (UniqueName: \"kubernetes.io/projected/43ead517-eb06-4b99-848f-7c49af380b6b-kube-api-access-4nnmv\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794266 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzscc\" (UniqueName: \"kubernetes.io/projected/53a59a18-ba76-4b57-bf78-d200d411f8b2-kube-api-access-hzscc\") pod \"kube-storage-version-migrator-operator-b67b599dd-4zwjj\" (UID: \"53a59a18-ba76-4b57-bf78-d200d411f8b2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794289 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/64014133-dfa9-4052-83a3-c88736be91dd-signing-cabundle\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794307 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6d68711-24e1-46d1-95bb-31974c24b299-metrics-certs\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794322 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3158f6f2-98f3-4eb0-97d7-83c68dfeb743-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6dnsm\" (UID: \"3158f6f2-98f3-4eb0-97d7-83c68dfeb743\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794338 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-encryption-config\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794377 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbvjv\" (UniqueName: \"kubernetes.io/projected/69d90cc8-5227-4516-9537-5d3b83f92dbd-kube-api-access-nbvjv\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794393 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828da568-03e9-42d6-883b-7bbb1d429825-config\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794407 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-service-ca\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794409 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7305c6-5d47-4c06-aaa9-b4308c578ae6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ktchr\" (UID: \"2b7305c6-5d47-4c06-aaa9-b4308c578ae6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794423 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjzxd\" (UniqueName: \"kubernetes.io/projected/69b46aab-afdf-4551-b0b4-b8d4f8f97831-kube-api-access-vjzxd\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794443 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfrtd\" (UniqueName: \"kubernetes.io/projected/f9430f70-1c4a-47af-813f-76079af84e5e-kube-api-access-sfrtd\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794458 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53a59a18-ba76-4b57-bf78-d200d411f8b2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4zwjj\" (UID: \"53a59a18-ba76-4b57-bf78-d200d411f8b2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794975 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-dir\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.795301 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/69d90cc8-5227-4516-9537-5d3b83f92dbd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.796064 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.796310 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/828da568-03e9-42d6-883b-7bbb1d429825-service-ca-bundle\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.796345 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-image-import-ca\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.796778 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.798043 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-audit\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.798137 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2946e237-1d94-4eaa-ba11-148add5a3843-serving-cert\") pod \"openshift-config-operator-7777fb866f-wltgm\" (UID: \"2946e237-1d94-4eaa-ba11-148add5a3843\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.798427 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-etcd-client\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.798661 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-config\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.798720 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/43ead517-eb06-4b99-848f-7c49af380b6b-node-pullsecrets\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.798953 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9a23de-9ba3-4581-aa91-dfb253372643-config\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.799423 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.799525 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-policies\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.799571 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2946e237-1d94-4eaa-ba11-148add5a3843-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wltgm\" (UID: \"2946e237-1d94-4eaa-ba11-148add5a3843\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.799644 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-trusted-ca-bundle\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.800187 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9e9a23de-9ba3-4581-aa91-dfb253372643-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.800371 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/828da568-03e9-42d6-883b-7bbb1d429825-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.800490 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dzb6r"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.800559 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.800632 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.800684 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.800923 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.801585 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9e9a23de-9ba3-4581-aa91-dfb253372643-images\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.801610 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.793846 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.794159 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.801864 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-oauth-config\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.802013 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-audit-policies\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.802347 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/43ead517-eb06-4b99-848f-7c49af380b6b-encryption-config\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.802699 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.802800 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-serving-cert\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.803042 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-etcd-serving-ca\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.803270 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-service-ca\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.803378 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828da568-03e9-42d6-883b-7bbb1d429825-config\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.803593 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.803764 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-serving-cert\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.803794 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.804010 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.804303 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85065f7f-4262-49e8-8f55-4df7edc2b8a7-serving-cert\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.804325 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-config\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.804418 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.804436 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.804737 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43ead517-eb06-4b99-848f-7c49af380b6b-config\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.805006 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.805163 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-oauth-serving-cert\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.805550 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b7305c6-5d47-4c06-aaa9-b4308c578ae6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ktchr\" (UID: \"2b7305c6-5d47-4c06-aaa9-b4308c578ae6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.805580 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5vvq"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.805615 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/69d90cc8-5227-4516-9537-5d3b83f92dbd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.806107 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.806616 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.806842 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-encryption-config\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.806892 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.807199 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/828da568-03e9-42d6-883b-7bbb1d429825-serving-cert\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.807876 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.808670 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43ead517-eb06-4b99-848f-7c49af380b6b-serving-cert\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.809039 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wltgm"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.809762 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/43ead517-eb06-4b99-848f-7c49af380b6b-etcd-client\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.810419 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.810555 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.811520 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.811903 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.812819 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e5cb325-16ec-41f9-8336-dcb96a7160cb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fw8v8\" (UID: \"7e5cb325-16ec-41f9-8336-dcb96a7160cb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.813053 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.814056 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6j74g"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.815052 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.816539 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t4l7k"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.817660 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.818677 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bpmbs"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.819585 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jq5cl"] Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.823965 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.843315 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.863192 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.883229 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.895777 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e6d68711-24e1-46d1-95bb-31974c24b299-stats-auth\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.895829 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j79qp\" (UniqueName: \"kubernetes.io/projected/64014133-dfa9-4052-83a3-c88736be91dd-kube-api-access-j79qp\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.895857 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6d68711-24e1-46d1-95bb-31974c24b299-service-ca-bundle\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.895882 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-942lj\" (UniqueName: \"kubernetes.io/projected/e6d68711-24e1-46d1-95bb-31974c24b299-kube-api-access-942lj\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.895906 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr2r8\" (UniqueName: \"kubernetes.io/projected/2c35f71e-388f-48ce-8e84-9d177a72978b-kube-api-access-wr2r8\") pod \"package-server-manager-789f6589d5-8zrtm\" (UID: \"2c35f71e-388f-48ce-8e84-9d177a72978b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.895931 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3158f6f2-98f3-4eb0-97d7-83c68dfeb743-config\") pod \"kube-controller-manager-operator-78b949d7b-6dnsm\" (UID: \"3158f6f2-98f3-4eb0-97d7-83c68dfeb743\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.895953 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a59a18-ba76-4b57-bf78-d200d411f8b2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4zwjj\" (UID: \"53a59a18-ba76-4b57-bf78-d200d411f8b2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.895988 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/64014133-dfa9-4052-83a3-c88736be91dd-signing-key\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896020 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c35f71e-388f-48ce-8e84-9d177a72978b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8zrtm\" (UID: \"2c35f71e-388f-48ce-8e84-9d177a72978b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896045 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2dff72ce-7157-47d7-b48a-cf7e4c779d90-bound-sa-token\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896077 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzscc\" (UniqueName: \"kubernetes.io/projected/53a59a18-ba76-4b57-bf78-d200d411f8b2-kube-api-access-hzscc\") pod \"kube-storage-version-migrator-operator-b67b599dd-4zwjj\" (UID: \"53a59a18-ba76-4b57-bf78-d200d411f8b2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896102 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/64014133-dfa9-4052-83a3-c88736be91dd-signing-cabundle\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896125 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6d68711-24e1-46d1-95bb-31974c24b299-metrics-certs\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896176 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3158f6f2-98f3-4eb0-97d7-83c68dfeb743-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6dnsm\" (UID: \"3158f6f2-98f3-4eb0-97d7-83c68dfeb743\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896216 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfrtd\" (UniqueName: \"kubernetes.io/projected/f9430f70-1c4a-47af-813f-76079af84e5e-kube-api-access-sfrtd\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896239 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53a59a18-ba76-4b57-bf78-d200d411f8b2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4zwjj\" (UID: \"53a59a18-ba76-4b57-bf78-d200d411f8b2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896271 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e6d68711-24e1-46d1-95bb-31974c24b299-default-certificate\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896296 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-client-ca\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896319 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3158f6f2-98f3-4eb0-97d7-83c68dfeb743-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6dnsm\" (UID: \"3158f6f2-98f3-4eb0-97d7-83c68dfeb743\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896374 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlgwk\" (UniqueName: \"kubernetes.io/projected/e3be465c-283a-40aa-b004-5049a363ce29-kube-api-access-nlgwk\") pod \"dns-operator-744455d44c-pmls5\" (UID: \"e3be465c-283a-40aa-b004-5049a363ce29\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896411 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2dff72ce-7157-47d7-b48a-cf7e4c779d90-metrics-tls\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896434 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2dff72ce-7157-47d7-b48a-cf7e4c779d90-trusted-ca\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896478 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-config\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896502 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9430f70-1c4a-47af-813f-76079af84e5e-serving-cert\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896525 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4zld\" (UniqueName: \"kubernetes.io/projected/e44b12c7-ec46-4a33-b9ac-87668a6465b5-kube-api-access-x4zld\") pod \"olm-operator-6b444d44fb-rzkmk\" (UID: \"e44b12c7-ec46-4a33-b9ac-87668a6465b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896586 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3be465c-283a-40aa-b004-5049a363ce29-metrics-tls\") pod \"dns-operator-744455d44c-pmls5\" (UID: \"e3be465c-283a-40aa-b004-5049a363ce29\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896609 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e44b12c7-ec46-4a33-b9ac-87668a6465b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-rzkmk\" (UID: \"e44b12c7-ec46-4a33-b9ac-87668a6465b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896638 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e44b12c7-ec46-4a33-b9ac-87668a6465b5-srv-cert\") pod \"olm-operator-6b444d44fb-rzkmk\" (UID: \"e44b12c7-ec46-4a33-b9ac-87668a6465b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.896660 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5999\" (UniqueName: \"kubernetes.io/projected/2dff72ce-7157-47d7-b48a-cf7e4c779d90-kube-api-access-m5999\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.897339 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-client-ca\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.898216 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-config\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.900392 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9430f70-1c4a-47af-813f-76079af84e5e-serving-cert\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.911787 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3be465c-283a-40aa-b004-5049a363ce29-metrics-tls\") pod \"dns-operator-744455d44c-pmls5\" (UID: \"e3be465c-283a-40aa-b004-5049a363ce29\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.914258 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.923190 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.943407 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.963095 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 15:28:24 crc kubenswrapper[4805]: I1128 15:28:24.983188 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.003597 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.024488 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.030758 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3158f6f2-98f3-4eb0-97d7-83c68dfeb743-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6dnsm\" (UID: \"3158f6f2-98f3-4eb0-97d7-83c68dfeb743\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.043249 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.063857 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.066700 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3158f6f2-98f3-4eb0-97d7-83c68dfeb743-config\") pod \"kube-controller-manager-operator-78b949d7b-6dnsm\" (UID: \"3158f6f2-98f3-4eb0-97d7-83c68dfeb743\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.103827 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.123680 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.143995 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.151839 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2dff72ce-7157-47d7-b48a-cf7e4c779d90-metrics-tls\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.171645 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.177875 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2dff72ce-7157-47d7-b48a-cf7e4c779d90-trusted-ca\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.185249 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.203552 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.203623 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.203679 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.203937 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.204148 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.226865 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.252239 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.263630 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.283467 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.303453 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.324029 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.330199 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53a59a18-ba76-4b57-bf78-d200d411f8b2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4zwjj\" (UID: \"53a59a18-ba76-4b57-bf78-d200d411f8b2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.343673 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.364140 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.367123 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53a59a18-ba76-4b57-bf78-d200d411f8b2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4zwjj\" (UID: \"53a59a18-ba76-4b57-bf78-d200d411f8b2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.383402 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.403433 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.423437 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.430314 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e6d68711-24e1-46d1-95bb-31974c24b299-stats-auth\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.444113 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.464468 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.483678 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.489878 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6d68711-24e1-46d1-95bb-31974c24b299-metrics-certs\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.510826 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.520075 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e6d68711-24e1-46d1-95bb-31974c24b299-default-certificate\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.524398 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.544581 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.547395 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6d68711-24e1-46d1-95bb-31974c24b299-service-ca-bundle\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.563438 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.583837 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.603447 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.624028 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.643348 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.662791 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.683676 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.704290 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.721620 4805 request.go:700] Waited for 1.010552839s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/secrets?fieldSelector=metadata.name%3Dservice-ca-operator-dockercfg-rg9jl&limit=500&resourceVersion=0 Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.723440 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.743461 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.763263 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.783184 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.804021 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.823901 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.831644 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e44b12c7-ec46-4a33-b9ac-87668a6465b5-srv-cert\") pod \"olm-operator-6b444d44fb-rzkmk\" (UID: \"e44b12c7-ec46-4a33-b9ac-87668a6465b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.843031 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.850773 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e44b12c7-ec46-4a33-b9ac-87668a6465b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-rzkmk\" (UID: \"e44b12c7-ec46-4a33-b9ac-87668a6465b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.864049 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.884437 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 15:28:25 crc kubenswrapper[4805]: E1128 15:28:25.896596 4805 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 28 15:28:25 crc kubenswrapper[4805]: E1128 15:28:25.896656 4805 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Nov 28 15:28:25 crc kubenswrapper[4805]: E1128 15:28:25.896674 4805 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Nov 28 15:28:25 crc kubenswrapper[4805]: E1128 15:28:25.896700 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c35f71e-388f-48ce-8e84-9d177a72978b-package-server-manager-serving-cert podName:2c35f71e-388f-48ce-8e84-9d177a72978b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:26.396676403 +0000 UTC m=+133.446467724 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/2c35f71e-388f-48ce-8e84-9d177a72978b-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-8zrtm" (UID: "2c35f71e-388f-48ce-8e84-9d177a72978b") : failed to sync secret cache: timed out waiting for the condition Nov 28 15:28:25 crc kubenswrapper[4805]: E1128 15:28:25.896731 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64014133-dfa9-4052-83a3-c88736be91dd-signing-key podName:64014133-dfa9-4052-83a3-c88736be91dd nodeName:}" failed. No retries permitted until 2025-11-28 15:28:26.396711544 +0000 UTC m=+133.446502855 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/64014133-dfa9-4052-83a3-c88736be91dd-signing-key") pod "service-ca-9c57cc56f-jq5cl" (UID: "64014133-dfa9-4052-83a3-c88736be91dd") : failed to sync secret cache: timed out waiting for the condition Nov 28 15:28:25 crc kubenswrapper[4805]: E1128 15:28:25.896777 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/64014133-dfa9-4052-83a3-c88736be91dd-signing-cabundle podName:64014133-dfa9-4052-83a3-c88736be91dd nodeName:}" failed. No retries permitted until 2025-11-28 15:28:26.396750945 +0000 UTC m=+133.446542286 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/64014133-dfa9-4052-83a3-c88736be91dd-signing-cabundle") pod "service-ca-9c57cc56f-jq5cl" (UID: "64014133-dfa9-4052-83a3-c88736be91dd") : failed to sync configmap cache: timed out waiting for the condition Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.904422 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.924059 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.944066 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.970940 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 15:28:25 crc kubenswrapper[4805]: I1128 15:28:25.984075 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.003964 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.023473 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.043689 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.063533 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.083403 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.104108 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.123940 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.143047 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.163726 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.183990 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.203301 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.223928 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.244825 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.284458 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.304558 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.323296 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.343937 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.364041 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.385287 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.403796 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.418323 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/64014133-dfa9-4052-83a3-c88736be91dd-signing-key\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.418396 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c35f71e-388f-48ce-8e84-9d177a72978b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8zrtm\" (UID: \"2c35f71e-388f-48ce-8e84-9d177a72978b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.418441 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/64014133-dfa9-4052-83a3-c88736be91dd-signing-cabundle\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.419453 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/64014133-dfa9-4052-83a3-c88736be91dd-signing-cabundle\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.424069 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.424760 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/64014133-dfa9-4052-83a3-c88736be91dd-signing-key\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.425305 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c35f71e-388f-48ce-8e84-9d177a72978b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8zrtm\" (UID: \"2c35f71e-388f-48ce-8e84-9d177a72978b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.444203 4805 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.464169 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.484060 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.503811 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.523495 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.544014 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.563969 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.584579 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.604093 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.643433 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnrsh\" (UniqueName: \"kubernetes.io/projected/85065f7f-4262-49e8-8f55-4df7edc2b8a7-kube-api-access-vnrsh\") pod \"controller-manager-879f6c89f-q6czs\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.661258 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f8xj\" (UniqueName: \"kubernetes.io/projected/7c4bc4c6-cc03-45be-b486-6f4cb080d7b0-kube-api-access-2f8xj\") pod \"apiserver-7bbb656c7d-wpvb4\" (UID: \"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.688936 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjcrj\" (UniqueName: \"kubernetes.io/projected/2946e237-1d94-4eaa-ba11-148add5a3843-kube-api-access-hjcrj\") pod \"openshift-config-operator-7777fb866f-wltgm\" (UID: \"2946e237-1d94-4eaa-ba11-148add5a3843\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.701489 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzzck\" (UniqueName: \"kubernetes.io/projected/2b7305c6-5d47-4c06-aaa9-b4308c578ae6-kube-api-access-xzzck\") pod \"openshift-apiserver-operator-796bbdcf4f-ktchr\" (UID: \"2b7305c6-5d47-4c06-aaa9-b4308c578ae6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.717432 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.722083 4805 request.go:700] Waited for 1.922448639s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication-operator/serviceaccounts/authentication-operator/token Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.729657 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/69d90cc8-5227-4516-9537-5d3b83f92dbd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.750586 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgtlb\" (UniqueName: \"kubernetes.io/projected/828da568-03e9-42d6-883b-7bbb1d429825-kube-api-access-qgtlb\") pod \"authentication-operator-69f744f599-v7h7c\" (UID: \"828da568-03e9-42d6-883b-7bbb1d429825\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.752190 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.762503 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nnmv\" (UniqueName: \"kubernetes.io/projected/43ead517-eb06-4b99-848f-7c49af380b6b-kube-api-access-4nnmv\") pod \"apiserver-76f77b778f-scnxw\" (UID: \"43ead517-eb06-4b99-848f-7c49af380b6b\") " pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.780081 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjzxd\" (UniqueName: \"kubernetes.io/projected/69b46aab-afdf-4551-b0b4-b8d4f8f97831-kube-api-access-vjzxd\") pod \"console-f9d7485db-7bprj\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.808098 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwwjw\" (UniqueName: \"kubernetes.io/projected/97ea3cfc-6d2a-417f-a397-34a5760d0a9e-kube-api-access-fwwjw\") pod \"downloads-7954f5f757-2b2hq\" (UID: \"97ea3cfc-6d2a-417f-a397-34a5760d0a9e\") " pod="openshift-console/downloads-7954f5f757-2b2hq" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.826076 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.831136 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6ncs\" (UniqueName: \"kubernetes.io/projected/9e9a23de-9ba3-4581-aa91-dfb253372643-kube-api-access-t6ncs\") pod \"machine-api-operator-5694c8668f-vjvw8\" (UID: \"9e9a23de-9ba3-4581-aa91-dfb253372643\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.835432 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.838509 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.861731 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvrhs\" (UniqueName: \"kubernetes.io/projected/aebbc881-3380-4df9-b082-f6e0d16b8f8e-kube-api-access-kvrhs\") pod \"oauth-openshift-558db77b4-jgwwr\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.861880 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.873658 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbvjv\" (UniqueName: \"kubernetes.io/projected/69d90cc8-5227-4516-9537-5d3b83f92dbd-kube-api-access-nbvjv\") pod \"cluster-image-registry-operator-dc59b4c8b-nhlk5\" (UID: \"69d90cc8-5227-4516-9537-5d3b83f92dbd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.873914 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.879930 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.882559 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84kzg\" (UniqueName: \"kubernetes.io/projected/7e5cb325-16ec-41f9-8336-dcb96a7160cb-kube-api-access-84kzg\") pod \"cluster-samples-operator-665b6dd947-fw8v8\" (UID: \"7e5cb325-16ec-41f9-8336-dcb96a7160cb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.888252 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2b2hq" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.910137 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j79qp\" (UniqueName: \"kubernetes.io/projected/64014133-dfa9-4052-83a3-c88736be91dd-kube-api-access-j79qp\") pod \"service-ca-9c57cc56f-jq5cl\" (UID: \"64014133-dfa9-4052-83a3-c88736be91dd\") " pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.928779 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr2r8\" (UniqueName: \"kubernetes.io/projected/2c35f71e-388f-48ce-8e84-9d177a72978b-kube-api-access-wr2r8\") pod \"package-server-manager-789f6589d5-8zrtm\" (UID: \"2c35f71e-388f-48ce-8e84-9d177a72978b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.937031 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-942lj\" (UniqueName: \"kubernetes.io/projected/e6d68711-24e1-46d1-95bb-31974c24b299-kube-api-access-942lj\") pod \"router-default-5444994796-lfq9l\" (UID: \"e6d68711-24e1-46d1-95bb-31974c24b299\") " pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.958968 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2dff72ce-7157-47d7-b48a-cf7e4c779d90-bound-sa-token\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.983299 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wltgm"] Nov 28 15:28:26 crc kubenswrapper[4805]: I1128 15:28:26.984843 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzscc\" (UniqueName: \"kubernetes.io/projected/53a59a18-ba76-4b57-bf78-d200d411f8b2-kube-api-access-hzscc\") pod \"kube-storage-version-migrator-operator-b67b599dd-4zwjj\" (UID: \"53a59a18-ba76-4b57-bf78-d200d411f8b2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.012174 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfrtd\" (UniqueName: \"kubernetes.io/projected/f9430f70-1c4a-47af-813f-76079af84e5e-kube-api-access-sfrtd\") pod \"route-controller-manager-6576b87f9c-dcr6x\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.018240 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3158f6f2-98f3-4eb0-97d7-83c68dfeb743-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6dnsm\" (UID: \"3158f6f2-98f3-4eb0-97d7-83c68dfeb743\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.019204 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.030732 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.032950 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" Nov 28 15:28:27 crc kubenswrapper[4805]: W1128 15:28:27.041709 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6d68711_24e1_46d1_95bb_31974c24b299.slice/crio-1f7f5b18129f7e3e53390010e988a82f6971ecf4b60e056dbae789d13426b794 WatchSource:0}: Error finding container 1f7f5b18129f7e3e53390010e988a82f6971ecf4b60e056dbae789d13426b794: Status 404 returned error can't find the container with id 1f7f5b18129f7e3e53390010e988a82f6971ecf4b60e056dbae789d13426b794 Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.044161 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlgwk\" (UniqueName: \"kubernetes.io/projected/e3be465c-283a-40aa-b004-5049a363ce29-kube-api-access-nlgwk\") pod \"dns-operator-744455d44c-pmls5\" (UID: \"e3be465c-283a-40aa-b004-5049a363ce29\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.066468 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4zld\" (UniqueName: \"kubernetes.io/projected/e44b12c7-ec46-4a33-b9ac-87668a6465b5-kube-api-access-x4zld\") pod \"olm-operator-6b444d44fb-rzkmk\" (UID: \"e44b12c7-ec46-4a33-b9ac-87668a6465b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.085852 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5999\" (UniqueName: \"kubernetes.io/projected/2dff72ce-7157-47d7-b48a-cf7e4c779d90-kube-api-access-m5999\") pod \"ingress-operator-5b745b69d9-68s7f\" (UID: \"2dff72ce-7157-47d7-b48a-cf7e4c779d90\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.092740 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.093329 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.099672 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.104103 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.123833 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.143957 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.156431 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.165899 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.185931 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.203105 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.212348 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.324910 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.326112 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vjvw8"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.340222 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v7h7c"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.421042 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q6czs"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.431449 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7bprj"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.440550 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2b2hq"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.451132 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-scnxw"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.547141 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jgwwr"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.548266 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.588424 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.588525 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.589032 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.589164 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.589263 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.589968 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0edd21b-3a3d-41c4-8e8b-af747bd69575-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.589999 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-trusted-ca\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.590032 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-certificates\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.590055 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-bound-sa-token\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.590101 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.590131 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0edd21b-3a3d-41c4-8e8b-af747bd69575-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.590147 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgvb8\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-kube-api-access-pgvb8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.590201 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-tls\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: E1128 15:28:27.590644 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.090632755 +0000 UTC m=+135.140424066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:27 crc kubenswrapper[4805]: W1128 15:28:27.596745 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c4bc4c6_cc03_45be_b486_6f4cb080d7b0.slice/crio-3701ffef25ebcd3c720b5969e98f1673c3d76a42f7d8aecbec7973dfc1caac95 WatchSource:0}: Error finding container 3701ffef25ebcd3c720b5969e98f1673c3d76a42f7d8aecbec7973dfc1caac95: Status 404 returned error can't find the container with id 3701ffef25ebcd3c720b5969e98f1673c3d76a42f7d8aecbec7973dfc1caac95 Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.614166 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8"] Nov 28 15:28:27 crc kubenswrapper[4805]: W1128 15:28:27.614743 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85065f7f_4262_49e8_8f55_4df7edc2b8a7.slice/crio-7143883eed036ca11fddf45a52e706f6f64cb834bbb9b760fbace5a791a45bea WatchSource:0}: Error finding container 7143883eed036ca11fddf45a52e706f6f64cb834bbb9b760fbace5a791a45bea: Status 404 returned error can't find the container with id 7143883eed036ca11fddf45a52e706f6f64cb834bbb9b760fbace5a791a45bea Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.616306 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm"] Nov 28 15:28:27 crc kubenswrapper[4805]: W1128 15:28:27.618278 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53a59a18_ba76_4b57_bf78_d200d411f8b2.slice/crio-db5c3cb1ad996201d62bc4f68166900edc3fb2f33830b8ec22f5b2389cf86dc9 WatchSource:0}: Error finding container db5c3cb1ad996201d62bc4f68166900edc3fb2f33830b8ec22f5b2389cf86dc9: Status 404 returned error can't find the container with id db5c3cb1ad996201d62bc4f68166900edc3fb2f33830b8ec22f5b2389cf86dc9 Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.669202 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.675856 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jq5cl"] Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691370 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691658 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4af1cce0-5f55-4683-b242-b987f554c0a1-serving-cert\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691719 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c31d9745-7e08-42e3-8fe4-a61624577979-tmpfs\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691740 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c31d9745-7e08-42e3-8fe4-a61624577979-webhook-cert\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691759 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5e4e8fb3-4188-4025-b6f6-107c37a91afa-etcd-ca\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691776 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/771d3766-40e1-40f6-8af7-4146f4928c85-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7cm6s\" (UID: \"771d3766-40e1-40f6-8af7-4146f4928c85\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691799 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2480de-74cf-47e4-8bc8-caed90699b00-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pldmd\" (UID: \"fa2480de-74cf-47e4-8bc8-caed90699b00\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691823 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8pmv\" (UniqueName: \"kubernetes.io/projected/ed210d08-27a0-4b04-aa18-1387ca89d44a-kube-api-access-d8pmv\") pod \"marketplace-operator-79b997595-dzb6r\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691858 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-config\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691879 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7e7fb9ae-4306-4a32-b805-a316427ae4f9-proxy-tls\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691923 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-tls\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691972 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4af1cce0-5f55-4683-b242-b987f554c0a1-trusted-ca\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.691991 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c31d9745-7e08-42e3-8fe4-a61624577979-apiservice-cert\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692010 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xpkz\" (UniqueName: \"kubernetes.io/projected/4ddd7ac2-635b-4fba-9765-a81039204b8f-kube-api-access-7xpkz\") pod \"collect-profiles-29405715-bnsqc\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692071 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-machine-approver-tls\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692138 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwlsb\" (UniqueName: \"kubernetes.io/projected/c31d9745-7e08-42e3-8fe4-a61624577979-kube-api-access-jwlsb\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692167 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc6km\" (UniqueName: \"kubernetes.io/projected/3f965da4-23aa-427c-9d3d-838d80ce014a-kube-api-access-hc6km\") pod \"catalog-operator-68c6474976-pqxhf\" (UID: \"3f965da4-23aa-427c-9d3d-838d80ce014a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692192 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ddd7ac2-635b-4fba-9765-a81039204b8f-config-volume\") pod \"collect-profiles-29405715-bnsqc\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692230 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dr6z\" (UniqueName: \"kubernetes.io/projected/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-kube-api-access-4dr6z\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692273 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-trusted-ca\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692302 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0edd21b-3a3d-41c4-8e8b-af747bd69575-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692380 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-certificates\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692407 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjpn2\" (UniqueName: \"kubernetes.io/projected/771d3766-40e1-40f6-8af7-4146f4928c85-kube-api-access-tjpn2\") pod \"machine-config-controller-84d6567774-7cm6s\" (UID: \"771d3766-40e1-40f6-8af7-4146f4928c85\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692450 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-bound-sa-token\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692506 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dzb6r\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692534 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f211d722-e4f3-4674-963c-d59dd6348af3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7b6jm\" (UID: \"f211d722-e4f3-4674-963c-d59dd6348af3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692576 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa2480de-74cf-47e4-8bc8-caed90699b00-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pldmd\" (UID: \"fa2480de-74cf-47e4-8bc8-caed90699b00\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692614 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s54k9\" (UniqueName: \"kubernetes.io/projected/b3af018b-ccf9-407c-947d-e8304d18fa3c-kube-api-access-s54k9\") pod \"migrator-59844c95c7-z4x7x\" (UID: \"b3af018b-ccf9-407c-947d-e8304d18fa3c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692640 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f211d722-e4f3-4674-963c-d59dd6348af3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7b6jm\" (UID: \"f211d722-e4f3-4674-963c-d59dd6348af3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692666 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f211d722-e4f3-4674-963c-d59dd6348af3-config\") pod \"kube-apiserver-operator-766d6c64bb-7b6jm\" (UID: \"f211d722-e4f3-4674-963c-d59dd6348af3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692714 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0edd21b-3a3d-41c4-8e8b-af747bd69575-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692742 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgvb8\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-kube-api-access-pgvb8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692785 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8fsv\" (UniqueName: \"kubernetes.io/projected/5e4e8fb3-4188-4025-b6f6-107c37a91afa-kube-api-access-b8fsv\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.692813 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa2480de-74cf-47e4-8bc8-caed90699b00-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pldmd\" (UID: \"fa2480de-74cf-47e4-8bc8-caed90699b00\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:27 crc kubenswrapper[4805]: E1128 15:28:27.693392 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.1933486 +0000 UTC m=+135.243139901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.694017 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c942827-8ca3-4cf1-8c9b-b07042d73d8b-config\") pod \"service-ca-operator-777779d784-bpkg4\" (UID: \"7c942827-8ca3-4cf1-8c9b-b07042d73d8b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.694061 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e524995-9d68-4c48-bbbf-b79ebe705f05-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5jbs\" (UID: \"8e524995-9d68-4c48-bbbf-b79ebe705f05\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.694083 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mfd8\" (UniqueName: \"kubernetes.io/projected/4af1cce0-5f55-4683-b242-b987f554c0a1-kube-api-access-2mfd8\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.694112 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c942827-8ca3-4cf1-8c9b-b07042d73d8b-serving-cert\") pod \"service-ca-operator-777779d784-bpkg4\" (UID: \"7c942827-8ca3-4cf1-8c9b-b07042d73d8b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.694142 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5b445254-64aa-45a3-955c-56381ef3fef0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tfb5c\" (UID: \"5b445254-64aa-45a3-955c-56381ef3fef0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.694159 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml2l8\" (UniqueName: \"kubernetes.io/projected/7c942827-8ca3-4cf1-8c9b-b07042d73d8b-kube-api-access-ml2l8\") pod \"service-ca-operator-777779d784-bpkg4\" (UID: \"7c942827-8ca3-4cf1-8c9b-b07042d73d8b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.694177 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af1cce0-5f55-4683-b242-b987f554c0a1-config\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.694195 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j49v5\" (UniqueName: \"kubernetes.io/projected/5b445254-64aa-45a3-955c-56381ef3fef0-kube-api-access-j49v5\") pod \"multus-admission-controller-857f4d67dd-tfb5c\" (UID: \"5b445254-64aa-45a3-955c-56381ef3fef0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.694256 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7e7fb9ae-4306-4a32-b805-a316427ae4f9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.695028 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-trusted-ca\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.697329 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcwqd\" (UniqueName: \"kubernetes.io/projected/8e524995-9d68-4c48-bbbf-b79ebe705f05-kube-api-access-dcwqd\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5jbs\" (UID: \"8e524995-9d68-4c48-bbbf-b79ebe705f05\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.697388 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3f965da4-23aa-427c-9d3d-838d80ce014a-srv-cert\") pod \"catalog-operator-68c6474976-pqxhf\" (UID: \"3f965da4-23aa-427c-9d3d-838d80ce014a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.697415 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dzb6r\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.697448 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ddd7ac2-635b-4fba-9765-a81039204b8f-secret-volume\") pod \"collect-profiles-29405715-bnsqc\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.697478 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/771d3766-40e1-40f6-8af7-4146f4928c85-proxy-tls\") pod \"machine-config-controller-84d6567774-7cm6s\" (UID: \"771d3766-40e1-40f6-8af7-4146f4928c85\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.697870 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-certificates\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.698127 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8bxc\" (UniqueName: \"kubernetes.io/projected/7e7fb9ae-4306-4a32-b805-a316427ae4f9-kube-api-access-d8bxc\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.698565 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0edd21b-3a3d-41c4-8e8b-af747bd69575-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.698577 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e4e8fb3-4188-4025-b6f6-107c37a91afa-serving-cert\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.698778 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e4e8fb3-4188-4025-b6f6-107c37a91afa-etcd-service-ca\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.698847 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e4e8fb3-4188-4025-b6f6-107c37a91afa-etcd-client\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.699060 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3f965da4-23aa-427c-9d3d-838d80ce014a-profile-collector-cert\") pod \"catalog-operator-68c6474976-pqxhf\" (UID: \"3f965da4-23aa-427c-9d3d-838d80ce014a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.699110 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e524995-9d68-4c48-bbbf-b79ebe705f05-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5jbs\" (UID: \"8e524995-9d68-4c48-bbbf-b79ebe705f05\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.699140 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e4e8fb3-4188-4025-b6f6-107c37a91afa-config\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.699817 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7e7fb9ae-4306-4a32-b805-a316427ae4f9-images\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.700078 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-auth-proxy-config\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.700575 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0edd21b-3a3d-41c4-8e8b-af747bd69575-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.704182 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-tls\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: W1128 15:28:27.727786 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64014133_dfa9_4052_83a3_c88736be91dd.slice/crio-6078cf51ff33621647a8e1b853f0314169ebc63a664b1537e7bf5cb563891d59 WatchSource:0}: Error finding container 6078cf51ff33621647a8e1b853f0314169ebc63a664b1537e7bf5cb563891d59: Status 404 returned error can't find the container with id 6078cf51ff33621647a8e1b853f0314169ebc63a664b1537e7bf5cb563891d59 Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.753235 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-bound-sa-token\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.770445 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgvb8\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-kube-api-access-pgvb8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.801993 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7e7fb9ae-4306-4a32-b805-a316427ae4f9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802060 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcwqd\" (UniqueName: \"kubernetes.io/projected/8e524995-9d68-4c48-bbbf-b79ebe705f05-kube-api-access-dcwqd\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5jbs\" (UID: \"8e524995-9d68-4c48-bbbf-b79ebe705f05\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802079 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3f965da4-23aa-427c-9d3d-838d80ce014a-srv-cert\") pod \"catalog-operator-68c6474976-pqxhf\" (UID: \"3f965da4-23aa-427c-9d3d-838d80ce014a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802095 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/771d3766-40e1-40f6-8af7-4146f4928c85-proxy-tls\") pod \"machine-config-controller-84d6567774-7cm6s\" (UID: \"771d3766-40e1-40f6-8af7-4146f4928c85\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802116 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4f10987d-9552-4178-98b0-496694b72649-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-lqdbd\" (UID: \"4f10987d-9552-4178-98b0-496694b72649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802135 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dzb6r\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802152 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ddd7ac2-635b-4fba-9765-a81039204b8f-secret-volume\") pod \"collect-profiles-29405715-bnsqc\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802185 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8bxc\" (UniqueName: \"kubernetes.io/projected/7e7fb9ae-4306-4a32-b805-a316427ae4f9-kube-api-access-d8bxc\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802203 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e4e8fb3-4188-4025-b6f6-107c37a91afa-etcd-client\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802234 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e4e8fb3-4188-4025-b6f6-107c37a91afa-serving-cert\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802248 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e4e8fb3-4188-4025-b6f6-107c37a91afa-etcd-service-ca\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802265 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3f965da4-23aa-427c-9d3d-838d80ce014a-profile-collector-cert\") pod \"catalog-operator-68c6474976-pqxhf\" (UID: \"3f965da4-23aa-427c-9d3d-838d80ce014a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802283 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e524995-9d68-4c48-bbbf-b79ebe705f05-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5jbs\" (UID: \"8e524995-9d68-4c48-bbbf-b79ebe705f05\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802300 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e4e8fb3-4188-4025-b6f6-107c37a91afa-config\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802333 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7e7fb9ae-4306-4a32-b805-a316427ae4f9-images\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802374 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-auth-proxy-config\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802429 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-mountpoint-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802451 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4af1cce0-5f55-4683-b242-b987f554c0a1-serving-cert\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802467 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjgkj\" (UniqueName: \"kubernetes.io/projected/05b562d8-9b14-45c7-99d3-f486a53e497a-kube-api-access-tjgkj\") pod \"ingress-canary-6j74g\" (UID: \"05b562d8-9b14-45c7-99d3-f486a53e497a\") " pod="openshift-ingress-canary/ingress-canary-6j74g" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802484 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-plugins-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802514 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c31d9745-7e08-42e3-8fe4-a61624577979-tmpfs\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802533 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c31d9745-7e08-42e3-8fe4-a61624577979-webhook-cert\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802550 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5e4e8fb3-4188-4025-b6f6-107c37a91afa-etcd-ca\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802565 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/771d3766-40e1-40f6-8af7-4146f4928c85-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7cm6s\" (UID: \"771d3766-40e1-40f6-8af7-4146f4928c85\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802581 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2480de-74cf-47e4-8bc8-caed90699b00-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pldmd\" (UID: \"fa2480de-74cf-47e4-8bc8-caed90699b00\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802605 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-csi-data-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802661 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8pmv\" (UniqueName: \"kubernetes.io/projected/ed210d08-27a0-4b04-aa18-1387ca89d44a-kube-api-access-d8pmv\") pod \"marketplace-operator-79b997595-dzb6r\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802678 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-config\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802693 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7e7fb9ae-4306-4a32-b805-a316427ae4f9-proxy-tls\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802718 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05b562d8-9b14-45c7-99d3-f486a53e497a-cert\") pod \"ingress-canary-6j74g\" (UID: \"05b562d8-9b14-45c7-99d3-f486a53e497a\") " pod="openshift-ingress-canary/ingress-canary-6j74g" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802737 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4af1cce0-5f55-4683-b242-b987f554c0a1-trusted-ca\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802752 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5763f340-9cb7-474a-b799-b9784efc6d9a-config-volume\") pod \"dns-default-bpmbs\" (UID: \"5763f340-9cb7-474a-b799-b9784efc6d9a\") " pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802768 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5763f340-9cb7-474a-b799-b9784efc6d9a-metrics-tls\") pod \"dns-default-bpmbs\" (UID: \"5763f340-9cb7-474a-b799-b9784efc6d9a\") " pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802806 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c31d9745-7e08-42e3-8fe4-a61624577979-apiservice-cert\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802821 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xpkz\" (UniqueName: \"kubernetes.io/projected/4ddd7ac2-635b-4fba-9765-a81039204b8f-kube-api-access-7xpkz\") pod \"collect-profiles-29405715-bnsqc\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802838 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vm6h\" (UniqueName: \"kubernetes.io/projected/8a09e9cb-8113-4382-8d66-10a28595ba07-kube-api-access-2vm6h\") pod \"machine-config-server-trwpw\" (UID: \"8a09e9cb-8113-4382-8d66-10a28595ba07\") " pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802866 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-machine-approver-tls\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802922 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwlsb\" (UniqueName: \"kubernetes.io/projected/c31d9745-7e08-42e3-8fe4-a61624577979-kube-api-access-jwlsb\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802939 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc6km\" (UniqueName: \"kubernetes.io/projected/3f965da4-23aa-427c-9d3d-838d80ce014a-kube-api-access-hc6km\") pod \"catalog-operator-68c6474976-pqxhf\" (UID: \"3f965da4-23aa-427c-9d3d-838d80ce014a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802954 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ddd7ac2-635b-4fba-9765-a81039204b8f-config-volume\") pod \"collect-profiles-29405715-bnsqc\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802969 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dr6z\" (UniqueName: \"kubernetes.io/projected/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-kube-api-access-4dr6z\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.802986 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppbfs\" (UniqueName: \"kubernetes.io/projected/e2692e90-917d-4467-825d-5e277ce862bc-kube-api-access-ppbfs\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803003 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8a09e9cb-8113-4382-8d66-10a28595ba07-node-bootstrap-token\") pod \"machine-config-server-trwpw\" (UID: \"8a09e9cb-8113-4382-8d66-10a28595ba07\") " pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803031 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjpn2\" (UniqueName: \"kubernetes.io/projected/771d3766-40e1-40f6-8af7-4146f4928c85-kube-api-access-tjpn2\") pod \"machine-config-controller-84d6567774-7cm6s\" (UID: \"771d3766-40e1-40f6-8af7-4146f4928c85\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803054 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqnkh\" (UniqueName: \"kubernetes.io/projected/5763f340-9cb7-474a-b799-b9784efc6d9a-kube-api-access-rqnkh\") pod \"dns-default-bpmbs\" (UID: \"5763f340-9cb7-474a-b799-b9784efc6d9a\") " pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803071 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-registration-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803098 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dzb6r\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803115 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f211d722-e4f3-4674-963c-d59dd6348af3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7b6jm\" (UID: \"f211d722-e4f3-4674-963c-d59dd6348af3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803131 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa2480de-74cf-47e4-8bc8-caed90699b00-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pldmd\" (UID: \"fa2480de-74cf-47e4-8bc8-caed90699b00\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803147 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8a09e9cb-8113-4382-8d66-10a28595ba07-certs\") pod \"machine-config-server-trwpw\" (UID: \"8a09e9cb-8113-4382-8d66-10a28595ba07\") " pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803160 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-socket-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803186 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f211d722-e4f3-4674-963c-d59dd6348af3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7b6jm\" (UID: \"f211d722-e4f3-4674-963c-d59dd6348af3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803204 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s54k9\" (UniqueName: \"kubernetes.io/projected/b3af018b-ccf9-407c-947d-e8304d18fa3c-kube-api-access-s54k9\") pod \"migrator-59844c95c7-z4x7x\" (UID: \"b3af018b-ccf9-407c-947d-e8304d18fa3c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803234 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803250 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f211d722-e4f3-4674-963c-d59dd6348af3-config\") pod \"kube-apiserver-operator-766d6c64bb-7b6jm\" (UID: \"f211d722-e4f3-4674-963c-d59dd6348af3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803308 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8fsv\" (UniqueName: \"kubernetes.io/projected/5e4e8fb3-4188-4025-b6f6-107c37a91afa-kube-api-access-b8fsv\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803328 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpm5k\" (UniqueName: \"kubernetes.io/projected/4f10987d-9552-4178-98b0-496694b72649-kube-api-access-jpm5k\") pod \"control-plane-machine-set-operator-78cbb6b69f-lqdbd\" (UID: \"4f10987d-9552-4178-98b0-496694b72649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803383 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa2480de-74cf-47e4-8bc8-caed90699b00-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pldmd\" (UID: \"fa2480de-74cf-47e4-8bc8-caed90699b00\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803447 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c942827-8ca3-4cf1-8c9b-b07042d73d8b-config\") pod \"service-ca-operator-777779d784-bpkg4\" (UID: \"7c942827-8ca3-4cf1-8c9b-b07042d73d8b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803462 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e524995-9d68-4c48-bbbf-b79ebe705f05-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5jbs\" (UID: \"8e524995-9d68-4c48-bbbf-b79ebe705f05\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803504 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mfd8\" (UniqueName: \"kubernetes.io/projected/4af1cce0-5f55-4683-b242-b987f554c0a1-kube-api-access-2mfd8\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803519 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c942827-8ca3-4cf1-8c9b-b07042d73d8b-serving-cert\") pod \"service-ca-operator-777779d784-bpkg4\" (UID: \"7c942827-8ca3-4cf1-8c9b-b07042d73d8b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803563 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af1cce0-5f55-4683-b242-b987f554c0a1-config\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803582 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5b445254-64aa-45a3-955c-56381ef3fef0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tfb5c\" (UID: \"5b445254-64aa-45a3-955c-56381ef3fef0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803599 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml2l8\" (UniqueName: \"kubernetes.io/projected/7c942827-8ca3-4cf1-8c9b-b07042d73d8b-kube-api-access-ml2l8\") pod \"service-ca-operator-777779d784-bpkg4\" (UID: \"7c942827-8ca3-4cf1-8c9b-b07042d73d8b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.803616 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j49v5\" (UniqueName: \"kubernetes.io/projected/5b445254-64aa-45a3-955c-56381ef3fef0-kube-api-access-j49v5\") pod \"multus-admission-controller-857f4d67dd-tfb5c\" (UID: \"5b445254-64aa-45a3-955c-56381ef3fef0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.805491 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7e7fb9ae-4306-4a32-b805-a316427ae4f9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: E1128 15:28:27.807082 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.307062998 +0000 UTC m=+135.356854309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.807115 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e4e8fb3-4188-4025-b6f6-107c37a91afa-etcd-service-ca\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.807145 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5e4e8fb3-4188-4025-b6f6-107c37a91afa-etcd-ca\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.807417 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/771d3766-40e1-40f6-8af7-4146f4928c85-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7cm6s\" (UID: \"771d3766-40e1-40f6-8af7-4146f4928c85\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.808305 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f211d722-e4f3-4674-963c-d59dd6348af3-config\") pod \"kube-apiserver-operator-766d6c64bb-7b6jm\" (UID: \"f211d722-e4f3-4674-963c-d59dd6348af3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.810597 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2480de-74cf-47e4-8bc8-caed90699b00-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pldmd\" (UID: \"fa2480de-74cf-47e4-8bc8-caed90699b00\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.812088 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4af1cce0-5f55-4683-b242-b987f554c0a1-trusted-ca\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.812225 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e524995-9d68-4c48-bbbf-b79ebe705f05-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5jbs\" (UID: \"8e524995-9d68-4c48-bbbf-b79ebe705f05\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.813113 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c942827-8ca3-4cf1-8c9b-b07042d73d8b-config\") pod \"service-ca-operator-777779d784-bpkg4\" (UID: \"7c942827-8ca3-4cf1-8c9b-b07042d73d8b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.813231 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ddd7ac2-635b-4fba-9765-a81039204b8f-config-volume\") pod \"collect-profiles-29405715-bnsqc\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.813388 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e4e8fb3-4188-4025-b6f6-107c37a91afa-config\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.813924 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7e7fb9ae-4306-4a32-b805-a316427ae4f9-images\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.815038 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-auth-proxy-config\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.815054 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-config\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.818047 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af1cce0-5f55-4683-b242-b987f554c0a1-config\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.818068 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dzb6r\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.819670 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c31d9745-7e08-42e3-8fe4-a61624577979-tmpfs\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.904199 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.904371 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpm5k\" (UniqueName: \"kubernetes.io/projected/4f10987d-9552-4178-98b0-496694b72649-kube-api-access-jpm5k\") pod \"control-plane-machine-set-operator-78cbb6b69f-lqdbd\" (UID: \"4f10987d-9552-4178-98b0-496694b72649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" Nov 28 15:28:27 crc kubenswrapper[4805]: E1128 15:28:27.904397 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.404350257 +0000 UTC m=+135.454141578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.904561 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4f10987d-9552-4178-98b0-496694b72649-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-lqdbd\" (UID: \"4f10987d-9552-4178-98b0-496694b72649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.904665 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-mountpoint-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.904704 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjgkj\" (UniqueName: \"kubernetes.io/projected/05b562d8-9b14-45c7-99d3-f486a53e497a-kube-api-access-tjgkj\") pod \"ingress-canary-6j74g\" (UID: \"05b562d8-9b14-45c7-99d3-f486a53e497a\") " pod="openshift-ingress-canary/ingress-canary-6j74g" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.904727 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-plugins-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.904765 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-csi-data-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.904805 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05b562d8-9b14-45c7-99d3-f486a53e497a-cert\") pod \"ingress-canary-6j74g\" (UID: \"05b562d8-9b14-45c7-99d3-f486a53e497a\") " pod="openshift-ingress-canary/ingress-canary-6j74g" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.904830 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5763f340-9cb7-474a-b799-b9784efc6d9a-config-volume\") pod \"dns-default-bpmbs\" (UID: \"5763f340-9cb7-474a-b799-b9784efc6d9a\") " pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.905615 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5763f340-9cb7-474a-b799-b9784efc6d9a-metrics-tls\") pod \"dns-default-bpmbs\" (UID: \"5763f340-9cb7-474a-b799-b9784efc6d9a\") " pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.905696 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-csi-data-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.905719 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vm6h\" (UniqueName: \"kubernetes.io/projected/8a09e9cb-8113-4382-8d66-10a28595ba07-kube-api-access-2vm6h\") pod \"machine-config-server-trwpw\" (UID: \"8a09e9cb-8113-4382-8d66-10a28595ba07\") " pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.905816 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppbfs\" (UniqueName: \"kubernetes.io/projected/e2692e90-917d-4467-825d-5e277ce862bc-kube-api-access-ppbfs\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.905855 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8a09e9cb-8113-4382-8d66-10a28595ba07-node-bootstrap-token\") pod \"machine-config-server-trwpw\" (UID: \"8a09e9cb-8113-4382-8d66-10a28595ba07\") " pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.905915 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqnkh\" (UniqueName: \"kubernetes.io/projected/5763f340-9cb7-474a-b799-b9784efc6d9a-kube-api-access-rqnkh\") pod \"dns-default-bpmbs\" (UID: \"5763f340-9cb7-474a-b799-b9784efc6d9a\") " pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.905940 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-registration-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.905981 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8a09e9cb-8113-4382-8d66-10a28595ba07-certs\") pod \"machine-config-server-trwpw\" (UID: \"8a09e9cb-8113-4382-8d66-10a28595ba07\") " pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.906006 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-socket-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.906065 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.906241 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5763f340-9cb7-474a-b799-b9784efc6d9a-config-volume\") pod \"dns-default-bpmbs\" (UID: \"5763f340-9cb7-474a-b799-b9784efc6d9a\") " pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:27 crc kubenswrapper[4805]: E1128 15:28:27.906472 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.406457038 +0000 UTC m=+135.456248399 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.906510 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-plugins-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.906644 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-socket-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.906645 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-registration-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.906690 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e2692e90-917d-4467-825d-5e277ce862bc-mountpoint-dir\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.910757 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-machine-approver-tls\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.912637 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e4e8fb3-4188-4025-b6f6-107c37a91afa-etcd-client\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.918830 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dzb6r\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.918987 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s54k9\" (UniqueName: \"kubernetes.io/projected/b3af018b-ccf9-407c-947d-e8304d18fa3c-kube-api-access-s54k9\") pod \"migrator-59844c95c7-z4x7x\" (UID: \"b3af018b-ccf9-407c-947d-e8304d18fa3c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.919087 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/771d3766-40e1-40f6-8af7-4146f4928c85-proxy-tls\") pod \"machine-config-controller-84d6567774-7cm6s\" (UID: \"771d3766-40e1-40f6-8af7-4146f4928c85\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.972320 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" event={"ID":"2b7305c6-5d47-4c06-aaa9-b4308c578ae6","Type":"ContainerStarted","Data":"80958b19137386eaf02e1de5260652acc3a32811fbcbd0c210ce79c3a877c896"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.973261 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" event={"ID":"aebbc881-3380-4df9-b082-f6e0d16b8f8e","Type":"ContainerStarted","Data":"e9392d43ac9ded6ef3b6a7c15d003fa17493e8a1600ecb3dcaeae3a93513a456"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.974545 4805 generic.go:334] "Generic (PLEG): container finished" podID="2946e237-1d94-4eaa-ba11-148add5a3843" containerID="89d10ba5723140509dfbf071e04476740e8fdde45176f4544d9434f45644facc" exitCode=0 Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.974581 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" event={"ID":"2946e237-1d94-4eaa-ba11-148add5a3843","Type":"ContainerDied","Data":"89d10ba5723140509dfbf071e04476740e8fdde45176f4544d9434f45644facc"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.974610 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" event={"ID":"2946e237-1d94-4eaa-ba11-148add5a3843","Type":"ContainerStarted","Data":"da998c13c62d53006f39854f58e5d7cdb9f6b21e089d8107f2b9c80f020beda5"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.975387 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" event={"ID":"69d90cc8-5227-4516-9537-5d3b83f92dbd","Type":"ContainerStarted","Data":"9135e54a2ab84bf6acc3524b9bd49407256c8e04fd15b8bdaf7b926f5a67f874"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.976048 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" event={"ID":"9e9a23de-9ba3-4581-aa91-dfb253372643","Type":"ContainerStarted","Data":"57f19efa6a9daf31f68501532486a9da578e351daae4109285ad3c8c0a11b38c"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.976920 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" event={"ID":"828da568-03e9-42d6-883b-7bbb1d429825","Type":"ContainerStarted","Data":"39e37d9deb0e1eefb4dcea2c93a14238e95eaaa2022d65a55a17b16b532e5e1d"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.977770 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7bprj" event={"ID":"69b46aab-afdf-4551-b0b4-b8d4f8f97831","Type":"ContainerStarted","Data":"c11e8a56bf01975264f72042e359a8966be236c441b3486b093aeef107642284"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.978436 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" event={"ID":"53a59a18-ba76-4b57-bf78-d200d411f8b2","Type":"ContainerStarted","Data":"db5c3cb1ad996201d62bc4f68166900edc3fb2f33830b8ec22f5b2389cf86dc9"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.979010 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" event={"ID":"2c35f71e-388f-48ce-8e84-9d177a72978b","Type":"ContainerStarted","Data":"4f576baebfe97d9c10fe0b1be06ddaf0969ca6eec89a4b667f0a566992400129"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.979679 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" event={"ID":"85065f7f-4262-49e8-8f55-4df7edc2b8a7","Type":"ContainerStarted","Data":"7143883eed036ca11fddf45a52e706f6f64cb834bbb9b760fbace5a791a45bea"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.980211 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2b2hq" event={"ID":"97ea3cfc-6d2a-417f-a397-34a5760d0a9e","Type":"ContainerStarted","Data":"b563b9a545953d45bb7380b7c3d529b6e237fab86a580f103bb1f823f952fe98"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.980778 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" event={"ID":"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0","Type":"ContainerStarted","Data":"3701ffef25ebcd3c720b5969e98f1673c3d76a42f7d8aecbec7973dfc1caac95"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.981346 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" event={"ID":"43ead517-eb06-4b99-848f-7c49af380b6b","Type":"ContainerStarted","Data":"fb3e68938738e1fe8d90e862bed3581e66bccc19a24f0acff27e3ad5a396e2f1"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.981973 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" event={"ID":"64014133-dfa9-4052-83a3-c88736be91dd","Type":"ContainerStarted","Data":"6078cf51ff33621647a8e1b853f0314169ebc63a664b1537e7bf5cb563891d59"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.982569 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" event={"ID":"7e5cb325-16ec-41f9-8336-dcb96a7160cb","Type":"ContainerStarted","Data":"7d0c5d78c2c8bf03e4920859df2a29c7f75c74b3f977088510fafacf4bfac628"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.983312 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lfq9l" event={"ID":"e6d68711-24e1-46d1-95bb-31974c24b299","Type":"ContainerStarted","Data":"f25561eca21eb97cf58f4e7d30217f783329ac84372706874a55a5cfe7c49067"} Nov 28 15:28:27 crc kubenswrapper[4805]: I1128 15:28:27.983335 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lfq9l" event={"ID":"e6d68711-24e1-46d1-95bb-31974c24b299","Type":"ContainerStarted","Data":"1f7f5b18129f7e3e53390010e988a82f6971ecf4b60e056dbae789d13426b794"} Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.007453 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.007832 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.507810943 +0000 UTC m=+135.557602254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.009885 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05b562d8-9b14-45c7-99d3-f486a53e497a-cert\") pod \"ingress-canary-6j74g\" (UID: \"05b562d8-9b14-45c7-99d3-f486a53e497a\") " pod="openshift-ingress-canary/ingress-canary-6j74g" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.009997 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8bxc\" (UniqueName: \"kubernetes.io/projected/7e7fb9ae-4306-4a32-b805-a316427ae4f9-kube-api-access-d8bxc\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.010384 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3f965da4-23aa-427c-9d3d-838d80ce014a-srv-cert\") pod \"catalog-operator-68c6474976-pqxhf\" (UID: \"3f965da4-23aa-427c-9d3d-838d80ce014a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.010384 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3f965da4-23aa-427c-9d3d-838d80ce014a-profile-collector-cert\") pod \"catalog-operator-68c6474976-pqxhf\" (UID: \"3f965da4-23aa-427c-9d3d-838d80ce014a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.010728 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f211d722-e4f3-4674-963c-d59dd6348af3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7b6jm\" (UID: \"f211d722-e4f3-4674-963c-d59dd6348af3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.011525 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7e7fb9ae-4306-4a32-b805-a316427ae4f9-proxy-tls\") pod \"machine-config-operator-74547568cd-fq7x2\" (UID: \"7e7fb9ae-4306-4a32-b805-a316427ae4f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.011680 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e4e8fb3-4188-4025-b6f6-107c37a91afa-serving-cert\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.011777 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c31d9745-7e08-42e3-8fe4-a61624577979-apiservice-cert\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.012014 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e524995-9d68-4c48-bbbf-b79ebe705f05-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5jbs\" (UID: \"8e524995-9d68-4c48-bbbf-b79ebe705f05\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.012266 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5763f340-9cb7-474a-b799-b9784efc6d9a-metrics-tls\") pod \"dns-default-bpmbs\" (UID: \"5763f340-9cb7-474a-b799-b9784efc6d9a\") " pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.012265 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8pmv\" (UniqueName: \"kubernetes.io/projected/ed210d08-27a0-4b04-aa18-1387ca89d44a-kube-api-access-d8pmv\") pod \"marketplace-operator-79b997595-dzb6r\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.013065 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4af1cce0-5f55-4683-b242-b987f554c0a1-serving-cert\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.014167 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f211d722-e4f3-4674-963c-d59dd6348af3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7b6jm\" (UID: \"f211d722-e4f3-4674-963c-d59dd6348af3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.016038 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcwqd\" (UniqueName: \"kubernetes.io/projected/8e524995-9d68-4c48-bbbf-b79ebe705f05-kube-api-access-dcwqd\") pod \"openshift-controller-manager-operator-756b6f6bc6-x5jbs\" (UID: \"8e524995-9d68-4c48-bbbf-b79ebe705f05\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.016426 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa2480de-74cf-47e4-8bc8-caed90699b00-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pldmd\" (UID: \"fa2480de-74cf-47e4-8bc8-caed90699b00\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.016764 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c31d9745-7e08-42e3-8fe4-a61624577979-webhook-cert\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.016954 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ddd7ac2-635b-4fba-9765-a81039204b8f-secret-volume\") pod \"collect-profiles-29405715-bnsqc\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.017061 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5b445254-64aa-45a3-955c-56381ef3fef0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tfb5c\" (UID: \"5b445254-64aa-45a3-955c-56381ef3fef0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.017653 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dr6z\" (UniqueName: \"kubernetes.io/projected/bec24d82-1d8f-44e2-b636-bfeb8d228ed5-kube-api-access-4dr6z\") pod \"machine-approver-56656f9798-26dv2\" (UID: \"bec24d82-1d8f-44e2-b636-bfeb8d228ed5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.019492 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.020621 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwlsb\" (UniqueName: \"kubernetes.io/projected/c31d9745-7e08-42e3-8fe4-a61624577979-kube-api-access-jwlsb\") pod \"packageserver-d55dfcdfc-qsq22\" (UID: \"c31d9745-7e08-42e3-8fe4-a61624577979\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.021093 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8a09e9cb-8113-4382-8d66-10a28595ba07-node-bootstrap-token\") pod \"machine-config-server-trwpw\" (UID: \"8a09e9cb-8113-4382-8d66-10a28595ba07\") " pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.032147 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa2480de-74cf-47e4-8bc8-caed90699b00-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pldmd\" (UID: \"fa2480de-74cf-47e4-8bc8-caed90699b00\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.032309 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4f10987d-9552-4178-98b0-496694b72649-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-lqdbd\" (UID: \"4f10987d-9552-4178-98b0-496694b72649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.032686 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjpn2\" (UniqueName: \"kubernetes.io/projected/771d3766-40e1-40f6-8af7-4146f4928c85-kube-api-access-tjpn2\") pod \"machine-config-controller-84d6567774-7cm6s\" (UID: \"771d3766-40e1-40f6-8af7-4146f4928c85\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.033058 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8a09e9cb-8113-4382-8d66-10a28595ba07-certs\") pod \"machine-config-server-trwpw\" (UID: \"8a09e9cb-8113-4382-8d66-10a28595ba07\") " pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.036892 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c942827-8ca3-4cf1-8c9b-b07042d73d8b-serving-cert\") pod \"service-ca-operator-777779d784-bpkg4\" (UID: \"7c942827-8ca3-4cf1-8c9b-b07042d73d8b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.055155 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j49v5\" (UniqueName: \"kubernetes.io/projected/5b445254-64aa-45a3-955c-56381ef3fef0-kube-api-access-j49v5\") pod \"multus-admission-controller-857f4d67dd-tfb5c\" (UID: \"5b445254-64aa-45a3-955c-56381ef3fef0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.058667 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:28 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:28 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:28 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.058736 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.071016 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xpkz\" (UniqueName: \"kubernetes.io/projected/4ddd7ac2-635b-4fba-9765-a81039204b8f-kube-api-access-7xpkz\") pod \"collect-profiles-29405715-bnsqc\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.092280 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc6km\" (UniqueName: \"kubernetes.io/projected/3f965da4-23aa-427c-9d3d-838d80ce014a-kube-api-access-hc6km\") pod \"catalog-operator-68c6474976-pqxhf\" (UID: \"3f965da4-23aa-427c-9d3d-838d80ce014a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.101648 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.102195 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mfd8\" (UniqueName: \"kubernetes.io/projected/4af1cce0-5f55-4683-b242-b987f554c0a1-kube-api-access-2mfd8\") pod \"console-operator-58897d9998-pzmqp\" (UID: \"4af1cce0-5f55-4683-b242-b987f554c0a1\") " pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.109701 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.110881 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.610865217 +0000 UTC m=+135.660656538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.128763 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8fsv\" (UniqueName: \"kubernetes.io/projected/5e4e8fb3-4188-4025-b6f6-107c37a91afa-kube-api-access-b8fsv\") pod \"etcd-operator-b45778765-vlp4p\" (UID: \"5e4e8fb3-4188-4025-b6f6-107c37a91afa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.138700 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.145080 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml2l8\" (UniqueName: \"kubernetes.io/projected/7c942827-8ca3-4cf1-8c9b-b07042d73d8b-kube-api-access-ml2l8\") pod \"service-ca-operator-777779d784-bpkg4\" (UID: \"7c942827-8ca3-4cf1-8c9b-b07042d73d8b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.146934 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.161735 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.193259 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.194313 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pmls5"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.199000 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpm5k\" (UniqueName: \"kubernetes.io/projected/4f10987d-9552-4178-98b0-496694b72649-kube-api-access-jpm5k\") pod \"control-plane-machine-set-operator-78cbb6b69f-lqdbd\" (UID: \"4f10987d-9552-4178-98b0-496694b72649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.201336 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.210602 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.212098 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.712076428 +0000 UTC m=+135.761867749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.215564 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjgkj\" (UniqueName: \"kubernetes.io/projected/05b562d8-9b14-45c7-99d3-f486a53e497a-kube-api-access-tjgkj\") pod \"ingress-canary-6j74g\" (UID: \"05b562d8-9b14-45c7-99d3-f486a53e497a\") " pod="openshift-ingress-canary/ingress-canary-6j74g" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.220426 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqnkh\" (UniqueName: \"kubernetes.io/projected/5763f340-9cb7-474a-b799-b9784efc6d9a-kube-api-access-rqnkh\") pod \"dns-default-bpmbs\" (UID: \"5763f340-9cb7-474a-b799-b9784efc6d9a\") " pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.227845 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.240544 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.246090 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.247501 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vm6h\" (UniqueName: \"kubernetes.io/projected/8a09e9cb-8113-4382-8d66-10a28595ba07-kube-api-access-2vm6h\") pod \"machine-config-server-trwpw\" (UID: \"8a09e9cb-8113-4382-8d66-10a28595ba07\") " pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.265235 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.267829 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.269917 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.278213 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.285154 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.294211 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppbfs\" (UniqueName: \"kubernetes.io/projected/e2692e90-917d-4467-825d-5e277ce862bc-kube-api-access-ppbfs\") pod \"csi-hostpathplugin-t4l7k\" (UID: \"e2692e90-917d-4467-825d-5e277ce862bc\") " pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.295138 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.308952 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.312351 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.312702 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.812687252 +0000 UTC m=+135.862478563 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.315739 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.321800 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.326459 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.328444 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.334030 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.367932 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.374533 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-trwpw" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.380550 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6j74g" Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.415104 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.415535 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:28.915504569 +0000 UTC m=+135.965295880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.480066 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.518826 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.519480 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.01946816 +0000 UTC m=+136.069259471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.621903 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.622105 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.12205636 +0000 UTC m=+136.171847671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.622688 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.623179 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.123154222 +0000 UTC m=+136.172945533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.730878 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.731141 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.231120876 +0000 UTC m=+136.280912187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.736968 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.737286 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.237273701 +0000 UTC m=+136.287065002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.791625 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vlp4p"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.796465 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pzmqp"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.842429 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.842929 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.342916889 +0000 UTC m=+136.392708200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.886598 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dzb6r"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.948138 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.949919 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:28 crc kubenswrapper[4805]: E1128 15:28:28.950604 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.450587715 +0000 UTC m=+136.500379026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.954522 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4"] Nov 28 15:28:28 crc kubenswrapper[4805]: I1128 15:28:28.985412 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.000766 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tfb5c"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.052860 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.053217 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.553200427 +0000 UTC m=+136.602991738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.075932 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-trwpw" event={"ID":"8a09e9cb-8113-4382-8d66-10a28595ba07","Type":"ContainerStarted","Data":"7fffe92af469f8be085ae97860378fb60cd86efd7000b38f659a7ea5efebec39"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.081457 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" event={"ID":"bec24d82-1d8f-44e2-b636-bfeb8d228ed5","Type":"ContainerStarted","Data":"ae6b56dcf51d0171292900bd1dc14642e1cd904b5a9d2ebaeba188fa810dd989"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.081494 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" event={"ID":"bec24d82-1d8f-44e2-b636-bfeb8d228ed5","Type":"ContainerStarted","Data":"f037f33d0bb6d65b15b0a63d4165f5c4d8323fc509ea796a94f1b1549d5b3572"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.097378 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" event={"ID":"2c35f71e-388f-48ce-8e84-9d177a72978b","Type":"ContainerStarted","Data":"80138180d389a3cf15bbd444e8346f36651c6a06c29239b6717de68db79d1955"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.109160 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" event={"ID":"e44b12c7-ec46-4a33-b9ac-87668a6465b5","Type":"ContainerStarted","Data":"0c91a1cc463f76a33a1bbb024e35d633a1e014079d4a2ee0c0564ea866729306"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.112072 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" event={"ID":"2dff72ce-7157-47d7-b48a-cf7e4c779d90","Type":"ContainerStarted","Data":"046263db1dc4b7b0adc0d2bd65e3ae9e289eed82865323c39c2f37e017f412f7"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.154813 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.155158 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.655116768 +0000 UTC m=+136.704908079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.165931 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:29 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:29 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:29 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.166026 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.169112 4805 generic.go:334] "Generic (PLEG): container finished" podID="43ead517-eb06-4b99-848f-7c49af380b6b" containerID="c0e038ee4ee32632802dff482190cb432cd6566820251cc9cd6e1e237f56ed15" exitCode=0 Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.169189 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" event={"ID":"43ead517-eb06-4b99-848f-7c49af380b6b","Type":"ContainerDied","Data":"c0e038ee4ee32632802dff482190cb432cd6566820251cc9cd6e1e237f56ed15"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.256891 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.257036 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.757011699 +0000 UTC m=+136.806803010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.257148 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.263758 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.763740781 +0000 UTC m=+136.813532082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.324886 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" event={"ID":"f9430f70-1c4a-47af-813f-76079af84e5e","Type":"ContainerStarted","Data":"d5a14efd4379f6d18bcea91e7410e861c9fda8fe6edea508196636661ffb4719"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.324932 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" event={"ID":"3158f6f2-98f3-4eb0-97d7-83c68dfeb743","Type":"ContainerStarted","Data":"5e419bbb4dd7b8a4aba13a0603503020fd3c6e25edab4851b6a507e2441bab50"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.324949 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7bprj" event={"ID":"69b46aab-afdf-4551-b0b4-b8d4f8f97831","Type":"ContainerStarted","Data":"ff69a3cd254e168cb937fb26fabceb62c0b5a467cb11b0339d2fc49212321969"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.324976 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc"] Nov 28 15:28:29 crc kubenswrapper[4805]: W1128 15:28:29.359468 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3af018b_ccf9_407c_947d_e8304d18fa3c.slice/crio-2dd345fae27afa1a5f2717f504a33b748dca53e1abda9e5ab6e3d420f1259b1b WatchSource:0}: Error finding container 2dd345fae27afa1a5f2717f504a33b748dca53e1abda9e5ab6e3d420f1259b1b: Status 404 returned error can't find the container with id 2dd345fae27afa1a5f2717f504a33b748dca53e1abda9e5ab6e3d420f1259b1b Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.360455 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-7bprj" podStartSLOduration=117.360439284 podStartE2EDuration="1m57.360439284s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.360075514 +0000 UTC m=+136.409866825" watchObservedRunningTime="2025-11-28 15:28:29.360439284 +0000 UTC m=+136.410230595" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.363306 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.364103 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.864078118 +0000 UTC m=+136.913869429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.366923 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.368948 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.868932607 +0000 UTC m=+136.918723918 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.395878 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" event={"ID":"9e9a23de-9ba3-4581-aa91-dfb253372643","Type":"ContainerStarted","Data":"d2322cddb66b821baf2c05683adf6c57c8a5df7edcf95351518f86725faf7ae2"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.405334 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bpmbs"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.421894 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lfq9l" podStartSLOduration=117.421873689 podStartE2EDuration="1m57.421873689s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.412855871 +0000 UTC m=+136.462647192" watchObservedRunningTime="2025-11-28 15:28:29.421873689 +0000 UTC m=+136.471665000" Nov 28 15:28:29 crc kubenswrapper[4805]: W1128 15:28:29.445300 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded210d08_27a0_4b04_aa18_1387ca89d44a.slice/crio-993142643a4512e108f050c2d339d7c8cc8661e3ec5b3dbf87a45ccd4bb30511 WatchSource:0}: Error finding container 993142643a4512e108f050c2d339d7c8cc8661e3ec5b3dbf87a45ccd4bb30511: Status 404 returned error can't find the container with id 993142643a4512e108f050c2d339d7c8cc8661e3ec5b3dbf87a45ccd4bb30511 Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.457172 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" event={"ID":"85065f7f-4262-49e8-8f55-4df7edc2b8a7","Type":"ContainerStarted","Data":"41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.460831 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.478163 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.478572 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:29.978557908 +0000 UTC m=+137.028349219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.490979 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.504533 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2b2hq" event={"ID":"97ea3cfc-6d2a-417f-a397-34a5760d0a9e","Type":"ContainerStarted","Data":"9bb8b904532ddd69011896cddad4b7d82a05c39f12b71947d9dc306b6d34ec8e"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.505244 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2b2hq" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.513728 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-2b2hq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.513789 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2b2hq" podUID="97ea3cfc-6d2a-417f-a397-34a5760d0a9e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.557249 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" event={"ID":"aebbc881-3380-4df9-b082-f6e0d16b8f8e","Type":"ContainerStarted","Data":"7b4663954d21b58e002fc6e305f94fd1b625d8112c41e6b5bfb0d34e95f6a00d"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.557313 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.583730 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.612195 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" event={"ID":"69d90cc8-5227-4516-9537-5d3b83f92dbd","Type":"ContainerStarted","Data":"53eb8a68bdf9f1d4c3d29b5bdad7248b13019e05f02f2c6155cdeba467ebacce"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.616604 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-2b2hq" podStartSLOduration=117.616588601 podStartE2EDuration="1m57.616588601s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.612178756 +0000 UTC m=+136.661970087" watchObservedRunningTime="2025-11-28 15:28:29.616588601 +0000 UTC m=+136.666379912" Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.635394 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.135375018 +0000 UTC m=+137.185166329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.636736 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" event={"ID":"64014133-dfa9-4052-83a3-c88736be91dd","Type":"ContainerStarted","Data":"ab0d51ed0b6f5443d81cf3c611ee2d511b6e714b2b94316fbdf75deffeece3cf"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.672442 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" event={"ID":"5e4e8fb3-4188-4025-b6f6-107c37a91afa","Type":"ContainerStarted","Data":"097e3396e909bd9700108df7706d7b1f44a16f6f95d08482c9fca4cbe0e543be"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.688036 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.690038 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.190012078 +0000 UTC m=+137.239803389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.695492 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" event={"ID":"2b7305c6-5d47-4c06-aaa9-b4308c578ae6","Type":"ContainerStarted","Data":"56e416d252c6c436fc3387a465dfcc46873766cc1936f690c8e7e4b9db29c9b9"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.714882 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" podStartSLOduration=117.714856848 podStartE2EDuration="1m57.714856848s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.552944043 +0000 UTC m=+136.602735344" watchObservedRunningTime="2025-11-28 15:28:29.714856848 +0000 UTC m=+136.764648149" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.736203 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.737285 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" podStartSLOduration=118.737275849 podStartE2EDuration="1m58.737275849s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.683114452 +0000 UTC m=+136.732905763" watchObservedRunningTime="2025-11-28 15:28:29.737275849 +0000 UTC m=+136.787067160" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.777543 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nhlk5" podStartSLOduration=117.777523279 podStartE2EDuration="1m57.777523279s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.717983307 +0000 UTC m=+136.767774618" watchObservedRunningTime="2025-11-28 15:28:29.777523279 +0000 UTC m=+136.827314590" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.780404 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.780439 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.781077 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" event={"ID":"7e5cb325-16ec-41f9-8336-dcb96a7160cb","Type":"ContainerStarted","Data":"68e2d7e3fba848d29cbbead8cda01d0cc56fb5d6ad190f1d18c6667863ca1d9a"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.792130 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.792746 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.292728533 +0000 UTC m=+137.342519844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.796288 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" event={"ID":"53a59a18-ba76-4b57-bf78-d200d411f8b2","Type":"ContainerStarted","Data":"6e36fb33c75bafbde404d0cb4c3cf0d398448dbc438be86b2cd2275af71912fb"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.827777 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-jq5cl" podStartSLOduration=117.827758294 podStartE2EDuration="1m57.827758294s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.770814257 +0000 UTC m=+136.820605568" watchObservedRunningTime="2025-11-28 15:28:29.827758294 +0000 UTC m=+136.877549605" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.829444 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.852985 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" event={"ID":"828da568-03e9-42d6-883b-7bbb1d429825","Type":"ContainerStarted","Data":"17a2740bee9f032bac0ae365febd85e281ee1177e9f012047ce1d26aecd3149e"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.860349 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t4l7k"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.861184 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ktchr" podStartSLOduration=118.861166418 podStartE2EDuration="1m58.861166418s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.806060044 +0000 UTC m=+136.855851365" watchObservedRunningTime="2025-11-28 15:28:29.861166418 +0000 UTC m=+136.910957729" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.863629 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4zwjj" podStartSLOduration=117.863620538 podStartE2EDuration="1m57.863620538s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.839322764 +0000 UTC m=+136.889114076" watchObservedRunningTime="2025-11-28 15:28:29.863620538 +0000 UTC m=+136.913411849" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.871633 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.890762 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7h7c" podStartSLOduration=118.890742933 podStartE2EDuration="1m58.890742933s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:29.88749242 +0000 UTC m=+136.937283731" watchObservedRunningTime="2025-11-28 15:28:29.890742933 +0000 UTC m=+136.940534244" Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.895523 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:29 crc kubenswrapper[4805]: E1128 15:28:29.897025 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.397005752 +0000 UTC m=+137.446797063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.899799 4805 generic.go:334] "Generic (PLEG): container finished" podID="7c4bc4c6-cc03-45be-b486-6f4cb080d7b0" containerID="fe335468ead614aea5c43e6cbf1b68f2072873d542c7bf126d28a457eaff0333" exitCode=0 Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.899909 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" event={"ID":"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0","Type":"ContainerDied","Data":"fe335468ead614aea5c43e6cbf1b68f2072873d542c7bf126d28a457eaff0333"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.931421 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" event={"ID":"8e524995-9d68-4c48-bbbf-b79ebe705f05","Type":"ContainerStarted","Data":"ce8015b8c427ebfca051e02a4e0ee8576824892f2bd9bed58bfc644aee87daec"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.955492 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" event={"ID":"e3be465c-283a-40aa-b004-5049a363ce29","Type":"ContainerStarted","Data":"6d8c384f366261e8b7b654b1e1d168a6ca46ba2888b6159b1060528c293c9305"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.967782 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s"] Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.987930 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" event={"ID":"2946e237-1d94-4eaa-ba11-148add5a3843","Type":"ContainerStarted","Data":"f70660e7306380df1ee60b4dd1939cce88fd4aec24a7986f7f8f689b78faaad8"} Nov 28 15:28:29 crc kubenswrapper[4805]: I1128 15:28:29.987971 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.014376 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.015978 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.515966181 +0000 UTC m=+137.565757492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.018696 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6j74g"] Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.018236 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" podStartSLOduration=119.018218485 podStartE2EDuration="1m59.018218485s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:30.016626709 +0000 UTC m=+137.066418020" watchObservedRunningTime="2025-11-28 15:28:30.018218485 +0000 UTC m=+137.068009796" Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.032081 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:30 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:30 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:30 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.032737 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:30 crc kubenswrapper[4805]: W1128 15:28:30.052485 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f965da4_23aa_427c_9d3d_838d80ce014a.slice/crio-b0438b8020c9fab3db4559b85653d199be7108581d41efdc7d0fec88b9cc162a WatchSource:0}: Error finding container b0438b8020c9fab3db4559b85653d199be7108581d41efdc7d0fec88b9cc162a: Status 404 returned error can't find the container with id b0438b8020c9fab3db4559b85653d199be7108581d41efdc7d0fec88b9cc162a Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.116978 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.123029 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.623012358 +0000 UTC m=+137.672803659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.220885 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.221260 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.721248725 +0000 UTC m=+137.771040046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.332695 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.332863 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.832838013 +0000 UTC m=+137.882629324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.333009 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.333323 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.833311236 +0000 UTC m=+137.883102547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.434276 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.434903 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:30.934890518 +0000 UTC m=+137.984681829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.539017 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.539383 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.039371643 +0000 UTC m=+138.089162954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.554974 4805 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jgwwr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.555028 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" podUID="aebbc881-3380-4df9-b082-f6e0d16b8f8e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.648815 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.649683 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.149663323 +0000 UTC m=+138.199454645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.762428 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.762712 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.262694012 +0000 UTC m=+138.312485323 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.866997 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.867698 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.367682462 +0000 UTC m=+138.417473773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:30 crc kubenswrapper[4805]: I1128 15:28:30.972022 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:30 crc kubenswrapper[4805]: E1128 15:28:30.972342 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.472328801 +0000 UTC m=+138.522120112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.030904 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:31 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:31 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:31 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.031390 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.073315 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:31 crc kubenswrapper[4805]: E1128 15:28:31.073652 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.573638415 +0000 UTC m=+138.623429716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.182042 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bpmbs" event={"ID":"5763f340-9cb7-474a-b799-b9784efc6d9a","Type":"ContainerStarted","Data":"eb6f947ee54cbb704ab0f2b692fdbf9d09c914917adcc311beaf5939ea5b54fb"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.182094 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bpmbs" event={"ID":"5763f340-9cb7-474a-b799-b9784efc6d9a","Type":"ContainerStarted","Data":"a265ca4ebddad710efd36e8c2c7e2f9ff0fa92ad403f11c25ffcb5d541449bd5"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.183584 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:31 crc kubenswrapper[4805]: E1128 15:28:31.183925 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.683910086 +0000 UTC m=+138.733701407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.236429 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" event={"ID":"3158f6f2-98f3-4eb0-97d7-83c68dfeb743","Type":"ContainerStarted","Data":"fc2da34e0c3152f98920c174c375a16fdd3af750ee8204e51e099620cd857d87"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.237845 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6dnsm" podStartSLOduration=119.237828716 podStartE2EDuration="1m59.237828716s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.236795357 +0000 UTC m=+138.286586668" watchObservedRunningTime="2025-11-28 15:28:31.237828716 +0000 UTC m=+138.287620027" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.240563 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" event={"ID":"4f10987d-9552-4178-98b0-496694b72649","Type":"ContainerStarted","Data":"b98934d41f7113149fa0f388676f3e640e8994aa0609bb713ed4050e99638eb1"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.240632 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" event={"ID":"4f10987d-9552-4178-98b0-496694b72649","Type":"ContainerStarted","Data":"441f8f9a6813d9dd38a5ebe1ebf153530200f6ed6b2022066c28881b7421cbf3"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.248072 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" event={"ID":"7c942827-8ca3-4cf1-8c9b-b07042d73d8b","Type":"ContainerStarted","Data":"bf5f681a1688f1eff588f2ea848b5e4f6710a979f4d546f1564f8fe7bdc3732d"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.248137 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" event={"ID":"7c942827-8ca3-4cf1-8c9b-b07042d73d8b","Type":"ContainerStarted","Data":"75f33523c9a98e71fd1aabd0c01f1364dac7e44aa849e837f8b9053a04aeaa4b"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.252198 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" event={"ID":"c31d9745-7e08-42e3-8fe4-a61624577979","Type":"ContainerStarted","Data":"10a17ef8f85aa790d084aa64b5c97c632d930c590776e0e82d8c04c3cc104443"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.252231 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" event={"ID":"c31d9745-7e08-42e3-8fe4-a61624577979","Type":"ContainerStarted","Data":"89ad558ed60806f0c54c2cdd7ec55fe50d6b0f5dc79f5eee1cdcf4c75a230384"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.253033 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.255290 4805 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-qsq22 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.255344 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" podUID="c31d9745-7e08-42e3-8fe4-a61624577979" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.255999 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6j74g" event={"ID":"05b562d8-9b14-45c7-99d3-f486a53e497a","Type":"ContainerStarted","Data":"630eb658717812611d88d7c71f6c39ae88f09f95240975282a47dfbaa5da3543"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.256034 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6j74g" event={"ID":"05b562d8-9b14-45c7-99d3-f486a53e497a","Type":"ContainerStarted","Data":"abce13bec7ddc15bfc60b87a3950a6a99770742d3989f9936888ca90d308a60a"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.257828 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" event={"ID":"f211d722-e4f3-4674-963c-d59dd6348af3","Type":"ContainerStarted","Data":"49b5d9ffe9249879bcbc0863f0fc581d330344f7e423129e1e8d6cc312902faa"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.259811 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" event={"ID":"e3be465c-283a-40aa-b004-5049a363ce29","Type":"ContainerStarted","Data":"0afce678e6e6c48889622126f7d6209d06b4958d930a006906eed7505254a62c"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.284562 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:31 crc kubenswrapper[4805]: E1128 15:28:31.285597 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.78558182 +0000 UTC m=+138.835373131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.292437 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lqdbd" podStartSLOduration=119.292421905 podStartE2EDuration="1m59.292421905s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.289936964 +0000 UTC m=+138.339728275" watchObservedRunningTime="2025-11-28 15:28:31.292421905 +0000 UTC m=+138.342213216" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.293158 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" event={"ID":"5e4e8fb3-4188-4025-b6f6-107c37a91afa","Type":"ContainerStarted","Data":"1edbfb9a1f52c9ddc677fa114181dc8ec3cbf001b99122b463b1271da2c06623"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.306577 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" event={"ID":"f9430f70-1c4a-47af-813f-76079af84e5e","Type":"ContainerStarted","Data":"4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.307510 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.318143 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-trwpw" event={"ID":"8a09e9cb-8113-4382-8d66-10a28595ba07","Type":"ContainerStarted","Data":"4f1f5923fc21fe20b549d969ce65424b4da843b35c6d7095e5ab8bd686790bfc"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.331545 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-6j74g" podStartSLOduration=7.331531452 podStartE2EDuration="7.331531452s" podCreationTimestamp="2025-11-28 15:28:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.32899279 +0000 UTC m=+138.378784101" watchObservedRunningTime="2025-11-28 15:28:31.331531452 +0000 UTC m=+138.381322763" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.342677 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.349760 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" event={"ID":"9e9a23de-9ba3-4581-aa91-dfb253372643","Type":"ContainerStarted","Data":"536d58b757aee8de82f91e6e3bb911f84ba9e5f58480910c127f8338783684f4"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.359759 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" event={"ID":"3f965da4-23aa-427c-9d3d-838d80ce014a","Type":"ContainerStarted","Data":"0d74c9ceb1da4da8deff91e5829be4caac27d17fc43a8a35d50275b908d12c9b"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.359808 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" event={"ID":"3f965da4-23aa-427c-9d3d-838d80ce014a","Type":"ContainerStarted","Data":"b0438b8020c9fab3db4559b85653d199be7108581d41efdc7d0fec88b9cc162a"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.360032 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bpkg4" podStartSLOduration=119.360015206 podStartE2EDuration="1m59.360015206s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.358007769 +0000 UTC m=+138.407799080" watchObservedRunningTime="2025-11-28 15:28:31.360015206 +0000 UTC m=+138.409806517" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.364587 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" event={"ID":"e44b12c7-ec46-4a33-b9ac-87668a6465b5","Type":"ContainerStarted","Data":"3bee31d9d14b18bab5d1382b49b018b3b55ac72a0144e84f958f5807dc11d868"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.364866 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.385771 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" podStartSLOduration=119.385754242 podStartE2EDuration="1m59.385754242s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.384222728 +0000 UTC m=+138.434014039" watchObservedRunningTime="2025-11-28 15:28:31.385754242 +0000 UTC m=+138.435545553" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.385938 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:31 crc kubenswrapper[4805]: E1128 15:28:31.386290 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.886274337 +0000 UTC m=+138.936065638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.390447 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.399612 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" event={"ID":"2c35f71e-388f-48ce-8e84-9d177a72978b","Type":"ContainerStarted","Data":"d555233a448f5d8a871d4f28a8239c5a657be1f455a7397a356910439bfd27b4"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.400242 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.414253 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" event={"ID":"2dff72ce-7157-47d7-b48a-cf7e4c779d90","Type":"ContainerStarted","Data":"146e4f9f40b80739ea7937ab837375c430938ad1d5c19b7766c51ab71b8df655"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.414299 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" event={"ID":"2dff72ce-7157-47d7-b48a-cf7e4c779d90","Type":"ContainerStarted","Data":"aa92c2f9e97e7a2fa7dd8ced8370184615414860aa73ab87a9be85bb10d34f21"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.414453 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" podStartSLOduration=119.414437931 podStartE2EDuration="1m59.414437931s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.412733422 +0000 UTC m=+138.462524733" watchObservedRunningTime="2025-11-28 15:28:31.414437931 +0000 UTC m=+138.464229232" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.432672 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x" event={"ID":"b3af018b-ccf9-407c-947d-e8304d18fa3c","Type":"ContainerStarted","Data":"c428c22f939759b1d428043502b413471e7ffa74a5b6f44db97cf2538051b4cf"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.432715 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x" event={"ID":"b3af018b-ccf9-407c-947d-e8304d18fa3c","Type":"ContainerStarted","Data":"2dd345fae27afa1a5f2717f504a33b748dca53e1abda9e5ab6e3d420f1259b1b"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.440851 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" event={"ID":"4ddd7ac2-635b-4fba-9765-a81039204b8f","Type":"ContainerStarted","Data":"a40ad42cb4c8682b53fd13bb455f69ec3749f5082fd18e9951e96e984877794c"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.440878 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" event={"ID":"4ddd7ac2-635b-4fba-9765-a81039204b8f","Type":"ContainerStarted","Data":"ab82356be96a522e8277bac4728b5a1aea4711adfa485fcd0569a8ed2618c607"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.459149 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rzkmk" podStartSLOduration=119.459128798 podStartE2EDuration="1m59.459128798s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.439815126 +0000 UTC m=+138.489606447" watchObservedRunningTime="2025-11-28 15:28:31.459128798 +0000 UTC m=+138.508920109" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.460581 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-vjvw8" podStartSLOduration=119.460574879 podStartE2EDuration="1m59.460574879s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.459471058 +0000 UTC m=+138.509262369" watchObservedRunningTime="2025-11-28 15:28:31.460574879 +0000 UTC m=+138.510366190" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.487801 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:31 crc kubenswrapper[4805]: E1128 15:28:31.488707 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:31.988692342 +0000 UTC m=+139.038483643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.513421 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" event={"ID":"7e5cb325-16ec-41f9-8336-dcb96a7160cb","Type":"ContainerStarted","Data":"9b495cd9aa424d09a9adeeba3359778d291e346e5e81e8c97d49263f9d3873cc"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.541483 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" event={"ID":"bec24d82-1d8f-44e2-b636-bfeb8d228ed5","Type":"ContainerStarted","Data":"a7a97b78862bc7273b972869ec79921d7524b92685f01d1dec262c027cf8aed1"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.553519 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pzmqp" event={"ID":"4af1cce0-5f55-4683-b242-b987f554c0a1","Type":"ContainerStarted","Data":"cdc8d21b567b79919783ba4bdf2c76f3414191557c658190570b0c1585289320"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.553574 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pzmqp" event={"ID":"4af1cce0-5f55-4683-b242-b987f554c0a1","Type":"ContainerStarted","Data":"fcff299191cbb3a66ba400572d5c661dc7a02529a1da3ab8b59d7a96f9972b28"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.554442 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.561735 4805 patch_prober.go:28] interesting pod/console-operator-58897d9998-pzmqp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.561794 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-pzmqp" podUID="4af1cce0-5f55-4683-b242-b987f554c0a1" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.566348 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" event={"ID":"e2692e90-917d-4467-825d-5e277ce862bc","Type":"ContainerStarted","Data":"8eec8ea3e262d2a98695c6b678780e5dbaba1cd86652d03cb30a572491165475"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.586203 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" event={"ID":"8e524995-9d68-4c48-bbbf-b79ebe705f05","Type":"ContainerStarted","Data":"bbb87604be0aa09cbc7ee36e0a61b68192a11bccaeca2f897f5f3b6af81ff530"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.590782 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:31 crc kubenswrapper[4805]: E1128 15:28:31.591700 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.091683394 +0000 UTC m=+139.141474705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.592193 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" event={"ID":"fa2480de-74cf-47e4-8bc8-caed90699b00","Type":"ContainerStarted","Data":"05eda87e6b87cb3a066c9997f9b917ae2671cbc3149a1cd4b09e6031607f17b9"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.600724 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" event={"ID":"5b445254-64aa-45a3-955c-56381ef3fef0","Type":"ContainerStarted","Data":"9264814cb693ff4264cabd096b6e2bf57e349efa2e6a7bd0be9d4c9bf53bd2c7"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.600786 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" event={"ID":"5b445254-64aa-45a3-955c-56381ef3fef0","Type":"ContainerStarted","Data":"b8bf308b0e4c5166ab94ba1717d2e468396c8162ac680a6c746480d2bfd64360"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.620433 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vlp4p" podStartSLOduration=119.620414765 podStartE2EDuration="1m59.620414765s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.577817598 +0000 UTC m=+138.627608909" watchObservedRunningTime="2025-11-28 15:28:31.620414765 +0000 UTC m=+138.670206076" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.621783 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-trwpw" podStartSLOduration=7.621775994 podStartE2EDuration="7.621775994s" podCreationTimestamp="2025-11-28 15:28:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.613088896 +0000 UTC m=+138.662880207" watchObservedRunningTime="2025-11-28 15:28:31.621775994 +0000 UTC m=+138.671567295" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.622588 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" event={"ID":"771d3766-40e1-40f6-8af7-4146f4928c85","Type":"ContainerStarted","Data":"46e7c46f14be56c52420d8437874b00897eef002699322e67bac28d399d4f0ce"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.622646 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" event={"ID":"771d3766-40e1-40f6-8af7-4146f4928c85","Type":"ContainerStarted","Data":"9bb1060729f5c6a67f903f7b4c63b8e60e34a19b24385415066ff147cbb93377"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.677733 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fw8v8" podStartSLOduration=120.677717262 podStartE2EDuration="2m0.677717262s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.676703754 +0000 UTC m=+138.726495065" watchObservedRunningTime="2025-11-28 15:28:31.677717262 +0000 UTC m=+138.727508573" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.680059 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" event={"ID":"43ead517-eb06-4b99-848f-7c49af380b6b","Type":"ContainerStarted","Data":"5cfeadebc82cac6234db6efb50003626bd7d824c1236e8df8467dfe2d43a66b1"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.693611 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:31 crc kubenswrapper[4805]: E1128 15:28:31.694652 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.194633836 +0000 UTC m=+139.244425157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.712015 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" podStartSLOduration=119.711996511 podStartE2EDuration="1m59.711996511s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.711982131 +0000 UTC m=+138.761773452" watchObservedRunningTime="2025-11-28 15:28:31.711996511 +0000 UTC m=+138.761787822" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.712243 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" event={"ID":"ed210d08-27a0-4b04-aa18-1387ca89d44a","Type":"ContainerStarted","Data":"783b3213f3e536db2bb1a0188bf5cc31bd1ebe24c4a18274ee10234a7c598baa"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.712289 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" event={"ID":"ed210d08-27a0-4b04-aa18-1387ca89d44a","Type":"ContainerStarted","Data":"993142643a4512e108f050c2d339d7c8cc8661e3ec5b3dbf87a45ccd4bb30511"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.713341 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.732914 4805 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dzb6r container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.733204 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" podUID="ed210d08-27a0-4b04-aa18-1387ca89d44a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.757247 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" event={"ID":"7e7fb9ae-4306-4a32-b805-a316427ae4f9","Type":"ContainerStarted","Data":"0adc1208f9c25f012120e7f9587a838b8b839777eaed08f5bd780873763aed01"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.757285 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" event={"ID":"7e7fb9ae-4306-4a32-b805-a316427ae4f9","Type":"ContainerStarted","Data":"ab4bf245f4913edd4de90d567a99c3463e6d4bcf3dc2f248f1d9918562b4fa7e"} Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.757953 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-2b2hq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.757986 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2b2hq" podUID="97ea3cfc-6d2a-417f-a397-34a5760d0a9e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.769331 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.777690 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wltgm" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.795965 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:31 crc kubenswrapper[4805]: E1128 15:28:31.796449 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.296434713 +0000 UTC m=+139.346226024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.897265 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:31 crc kubenswrapper[4805]: E1128 15:28:31.897535 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.397517281 +0000 UTC m=+139.447308592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.910693 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26dv2" podStartSLOduration=120.910679347 podStartE2EDuration="2m0.910679347s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.906579211 +0000 UTC m=+138.956370522" watchObservedRunningTime="2025-11-28 15:28:31.910679347 +0000 UTC m=+138.960470658" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.911335 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-68s7f" podStartSLOduration=119.911329026 podStartE2EDuration="1m59.911329026s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.847692148 +0000 UTC m=+138.897483459" watchObservedRunningTime="2025-11-28 15:28:31.911329026 +0000 UTC m=+138.961120337" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.950628 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" podStartSLOduration=120.950609328 podStartE2EDuration="2m0.950609328s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.948792196 +0000 UTC m=+138.998583497" watchObservedRunningTime="2025-11-28 15:28:31.950609328 +0000 UTC m=+139.000400639" Nov 28 15:28:31 crc kubenswrapper[4805]: I1128 15:28:31.983775 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x5jbs" podStartSLOduration=119.983754185 podStartE2EDuration="1m59.983754185s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:31.98215683 +0000 UTC m=+139.031948141" watchObservedRunningTime="2025-11-28 15:28:31.983754185 +0000 UTC m=+139.033545486" Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.000993 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.001285 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.501274265 +0000 UTC m=+139.551065576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.023619 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-pzmqp" podStartSLOduration=120.023604703 podStartE2EDuration="2m0.023604703s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:32.022662486 +0000 UTC m=+139.072453797" watchObservedRunningTime="2025-11-28 15:28:32.023604703 +0000 UTC m=+139.073396014" Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.028527 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:32 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:32 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:32 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.028594 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.049704 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" podStartSLOduration=120.049682649 podStartE2EDuration="2m0.049682649s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:32.048733381 +0000 UTC m=+139.098524692" watchObservedRunningTime="2025-11-28 15:28:32.049682649 +0000 UTC m=+139.099473960" Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.102506 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.102612 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.60258198 +0000 UTC m=+139.652373291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.103905 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.104237 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.604229886 +0000 UTC m=+139.654021197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.204913 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.205101 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.705071347 +0000 UTC m=+139.754862678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.205176 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.205576 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.705564601 +0000 UTC m=+139.755355912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.305921 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.306090 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.806066622 +0000 UTC m=+139.855857933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.306215 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.306526 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.806514855 +0000 UTC m=+139.856306166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.407046 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.407197 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.90716878 +0000 UTC m=+139.956960091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.407379 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.407790 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:32.907778739 +0000 UTC m=+139.957570110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.508506 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.508965 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.008951129 +0000 UTC m=+140.058742440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.610111 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.610430 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.110417837 +0000 UTC m=+140.160209148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.711419 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.711603 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.211574607 +0000 UTC m=+140.261365918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.711781 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.712112 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.212101052 +0000 UTC m=+140.261892353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.762440 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" event={"ID":"fa2480de-74cf-47e4-8bc8-caed90699b00","Type":"ContainerStarted","Data":"1acc8d60063e4cfe7b2115758d095af582a74d852395aad85c612bb328c0a2f6"} Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.768324 4805 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dzb6r container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.768399 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" podUID="ed210d08-27a0-4b04-aa18-1387ca89d44a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.772543 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-2b2hq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.772621 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2b2hq" podUID="97ea3cfc-6d2a-417f-a397-34a5760d0a9e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.796562 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-pzmqp" Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.812866 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.813255 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.313225741 +0000 UTC m=+140.363017052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:32 crc kubenswrapper[4805]: I1128 15:28:32.914455 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:32 crc kubenswrapper[4805]: E1128 15:28:32.915529 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.415514703 +0000 UTC m=+140.465306014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.021971 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.022170 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.522145939 +0000 UTC m=+140.571937250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.022536 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.022809 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.522797508 +0000 UTC m=+140.572588819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.022892 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:33 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:33 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:33 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.022931 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.123234 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.123555 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.623541356 +0000 UTC m=+140.673332667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.224468 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.224878 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.72486606 +0000 UTC m=+140.774657361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.327424 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.327540 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.827518343 +0000 UTC m=+140.877309654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.328961 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.828941323 +0000 UTC m=+140.878732634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.328513 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.404678 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qsq22" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.431068 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.431244 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.931220035 +0000 UTC m=+140.981011346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.431348 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.431838 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:33.931831123 +0000 UTC m=+140.981622424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.537802 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.538337 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.038321205 +0000 UTC m=+141.088112506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.639311 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.639712 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.1396869 +0000 UTC m=+141.189478211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.740875 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.741055 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.241028435 +0000 UTC m=+141.290819746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.741235 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.741615 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.241599822 +0000 UTC m=+141.291391134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.767901 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" event={"ID":"e2692e90-917d-4467-825d-5e277ce862bc","Type":"ContainerStarted","Data":"c40f7ddc142738111581f2843ea9892abd567225d0b49fa3ee3a5bbefa231cc8"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.769465 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" event={"ID":"e3be465c-283a-40aa-b004-5049a363ce29","Type":"ContainerStarted","Data":"61db8a7937f989e72a157c9281a7daddec392a6ff78ad983e41f4ac5f3d826dc"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.771532 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bpmbs" event={"ID":"5763f340-9cb7-474a-b799-b9784efc6d9a","Type":"ContainerStarted","Data":"9698657c528ac99c220e2cd97678cab79d5f1879720595114c375b75393ee0bc"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.771915 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.773102 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" event={"ID":"7e7fb9ae-4306-4a32-b805-a316427ae4f9","Type":"ContainerStarted","Data":"0ffe2ee876c672cd9185f3099b7912c01de3156f1027d2dfd3cfadad1c5c3875"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.774247 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" event={"ID":"f211d722-e4f3-4674-963c-d59dd6348af3","Type":"ContainerStarted","Data":"8de74ce48fb2b3920f00e6c46a4972065d627d4c24fd33f0aefe03817816d28f"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.792585 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" event={"ID":"5b445254-64aa-45a3-955c-56381ef3fef0","Type":"ContainerStarted","Data":"fa01263eafa520d435f2d93340d39de86cbda1c582a2cde4abebe61da8566816"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.795486 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-pmls5" podStartSLOduration=121.795475241 podStartE2EDuration="2m1.795475241s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:33.795299246 +0000 UTC m=+140.845090557" watchObservedRunningTime="2025-11-28 15:28:33.795475241 +0000 UTC m=+140.845266542" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.805063 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" event={"ID":"771d3766-40e1-40f6-8af7-4146f4928c85","Type":"ContainerStarted","Data":"c472b04c36636f6e01519b10d1f7fe34f48ba7f166e59846c7107bbfbe7f0bb8"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.808415 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" event={"ID":"43ead517-eb06-4b99-848f-7c49af380b6b","Type":"ContainerStarted","Data":"04291b6fccca5b65f95a156c742cde1317579691c0d6c96299dc1c0b5cf85ebc"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.811745 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x" event={"ID":"b3af018b-ccf9-407c-947d-e8304d18fa3c","Type":"ContainerStarted","Data":"e6be8ac858f6d7a3dcea781d7d7767f793d2b0d76b7c7de476e6ceb5b46b34b7"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.814885 4805 generic.go:334] "Generic (PLEG): container finished" podID="4ddd7ac2-635b-4fba-9765-a81039204b8f" containerID="a40ad42cb4c8682b53fd13bb455f69ec3749f5082fd18e9951e96e984877794c" exitCode=0 Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.814950 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" event={"ID":"4ddd7ac2-635b-4fba-9765-a81039204b8f","Type":"ContainerDied","Data":"a40ad42cb4c8682b53fd13bb455f69ec3749f5082fd18e9951e96e984877794c"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.817733 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" event={"ID":"7c4bc4c6-cc03-45be-b486-6f4cb080d7b0","Type":"ContainerStarted","Data":"f57f95b36b26ff413d6b809c0a98546e30c0e1164f8b1bd465111b417521cc3e"} Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.825059 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-bpmbs" podStartSLOduration=9.825040265 podStartE2EDuration="9.825040265s" podCreationTimestamp="2025-11-28 15:28:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:33.823382269 +0000 UTC m=+140.873173580" watchObservedRunningTime="2025-11-28 15:28:33.825040265 +0000 UTC m=+140.874831566" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.843700 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.844184 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.844559 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-tfb5c" podStartSLOduration=121.844537862 podStartE2EDuration="2m1.844537862s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:33.842793583 +0000 UTC m=+140.892584894" watchObservedRunningTime="2025-11-28 15:28:33.844537862 +0000 UTC m=+140.894329173" Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.845110 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.345095389 +0000 UTC m=+141.394886700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.868781 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7b6jm" podStartSLOduration=121.868765645 podStartE2EDuration="2m1.868765645s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:33.865646466 +0000 UTC m=+140.915437787" watchObservedRunningTime="2025-11-28 15:28:33.868765645 +0000 UTC m=+140.918556956" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.896666 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fq7x2" podStartSLOduration=121.896653352 podStartE2EDuration="2m1.896653352s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:33.894762007 +0000 UTC m=+140.944553308" watchObservedRunningTime="2025-11-28 15:28:33.896653352 +0000 UTC m=+140.946444663" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.941413 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7cm6s" podStartSLOduration=121.94139193 podStartE2EDuration="2m1.94139193s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:33.933796762 +0000 UTC m=+140.983588073" watchObservedRunningTime="2025-11-28 15:28:33.94139193 +0000 UTC m=+140.991183241" Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.946632 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:33 crc kubenswrapper[4805]: E1128 15:28:33.953001 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.452988761 +0000 UTC m=+141.502780072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:33 crc kubenswrapper[4805]: I1128 15:28:33.958223 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" podStartSLOduration=121.95820416 podStartE2EDuration="2m1.95820416s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:33.957921552 +0000 UTC m=+141.007712863" watchObservedRunningTime="2025-11-28 15:28:33.95820416 +0000 UTC m=+141.007995471" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.024794 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" podStartSLOduration=122.024779622 podStartE2EDuration="2m2.024779622s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:34.024583987 +0000 UTC m=+141.074375318" watchObservedRunningTime="2025-11-28 15:28:34.024779622 +0000 UTC m=+141.074570933" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.029367 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:34 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:34 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:34 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.029416 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.043745 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pldmd" podStartSLOduration=122.043726643 podStartE2EDuration="2m2.043726643s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:34.041273073 +0000 UTC m=+141.091064384" watchObservedRunningTime="2025-11-28 15:28:34.043726643 +0000 UTC m=+141.093517954" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.049774 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.050116 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.550104716 +0000 UTC m=+141.599896027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.071768 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z4x7x" podStartSLOduration=122.071754993 podStartE2EDuration="2m2.071754993s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:34.064678882 +0000 UTC m=+141.114470193" watchObservedRunningTime="2025-11-28 15:28:34.071754993 +0000 UTC m=+141.121546304" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.073415 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8d4l4"] Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.074423 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.077404 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.100466 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" podStartSLOduration=123.100444794 podStartE2EDuration="2m3.100444794s" podCreationTimestamp="2025-11-28 15:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:34.100055003 +0000 UTC m=+141.149846314" watchObservedRunningTime="2025-11-28 15:28:34.100444794 +0000 UTC m=+141.150236105" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.115457 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8d4l4"] Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.152015 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.152063 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-utilities\") pod \"community-operators-8d4l4\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.152139 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phqdp\" (UniqueName: \"kubernetes.io/projected/db028335-3e13-44c6-844f-318168230b30-kube-api-access-phqdp\") pod \"community-operators-8d4l4\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.152165 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-catalog-content\") pod \"community-operators-8d4l4\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.152495 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.65248382 +0000 UTC m=+141.702275131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.252733 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dhzqc"] Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.253290 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.253527 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-utilities\") pod \"community-operators-8d4l4\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.253617 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phqdp\" (UniqueName: \"kubernetes.io/projected/db028335-3e13-44c6-844f-318168230b30-kube-api-access-phqdp\") pod \"community-operators-8d4l4\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.253654 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.253660 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-catalog-content\") pod \"community-operators-8d4l4\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.254084 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-catalog-content\") pod \"community-operators-8d4l4\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.254242 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.754220227 +0000 UTC m=+141.804011538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.254414 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-utilities\") pod \"community-operators-8d4l4\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.255989 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.262620 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dhzqc"] Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.275609 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phqdp\" (UniqueName: \"kubernetes.io/projected/db028335-3e13-44c6-844f-318168230b30-kube-api-access-phqdp\") pod \"community-operators-8d4l4\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.354694 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-catalog-content\") pod \"certified-operators-dhzqc\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.354924 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gssxt\" (UniqueName: \"kubernetes.io/projected/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-kube-api-access-gssxt\") pod \"certified-operators-dhzqc\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.355030 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-utilities\") pod \"certified-operators-dhzqc\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.355087 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.355397 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.855385547 +0000 UTC m=+141.905176858 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.403036 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.456687 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.95666717 +0000 UTC m=+142.006458481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.456684 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.456896 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-utilities\") pod \"certified-operators-dhzqc\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.456934 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.456986 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-catalog-content\") pod \"certified-operators-dhzqc\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.457067 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gssxt\" (UniqueName: \"kubernetes.io/projected/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-kube-api-access-gssxt\") pod \"certified-operators-dhzqc\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.457844 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-utilities\") pod \"certified-operators-dhzqc\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.458137 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:34.958125071 +0000 UTC m=+142.007916382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.469502 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rgs2f"] Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.472888 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.483232 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rgs2f"] Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.557561 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.557925 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-catalog-content\") pod \"community-operators-rgs2f\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.558022 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw2q2\" (UniqueName: \"kubernetes.io/projected/c6b04434-a30b-4e22-8c5a-9cdaca124466-kube-api-access-kw2q2\") pod \"community-operators-rgs2f\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.558054 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-utilities\") pod \"community-operators-rgs2f\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.558177 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.058163899 +0000 UTC m=+142.107955210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.608942 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8d4l4"] Nov 28 15:28:34 crc kubenswrapper[4805]: W1128 15:28:34.620999 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb028335_3e13_44c6_844f_318168230b30.slice/crio-e40e55c25a4094af294abe198a08c4afe5a2bd3f4829dc3c3a70182d263102d4 WatchSource:0}: Error finding container e40e55c25a4094af294abe198a08c4afe5a2bd3f4829dc3c3a70182d263102d4: Status 404 returned error can't find the container with id e40e55c25a4094af294abe198a08c4afe5a2bd3f4829dc3c3a70182d263102d4 Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.658919 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-catalog-content\") pod \"community-operators-rgs2f\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.659209 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw2q2\" (UniqueName: \"kubernetes.io/projected/c6b04434-a30b-4e22-8c5a-9cdaca124466-kube-api-access-kw2q2\") pod \"community-operators-rgs2f\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.659239 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-utilities\") pod \"community-operators-rgs2f\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.659260 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.659688 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.159659638 +0000 UTC m=+142.209450949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.658988 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7cqfs"] Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.660110 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-catalog-content\") pod \"community-operators-rgs2f\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.660562 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-utilities\") pod \"community-operators-rgs2f\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.661322 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.662704 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7cqfs"] Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.683804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw2q2\" (UniqueName: \"kubernetes.io/projected/c6b04434-a30b-4e22-8c5a-9cdaca124466-kube-api-access-kw2q2\") pod \"community-operators-rgs2f\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.760337 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.760520 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.260492679 +0000 UTC m=+142.310283990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.760651 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.760687 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-utilities\") pod \"certified-operators-7cqfs\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.760732 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mlsq\" (UniqueName: \"kubernetes.io/projected/d53216ca-283e-4346-ace1-9fab11518bfd-kube-api-access-7mlsq\") pod \"certified-operators-7cqfs\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.760757 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-catalog-content\") pod \"certified-operators-7cqfs\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.760995 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.260984534 +0000 UTC m=+142.310775845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.799815 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.832895 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d4l4" event={"ID":"db028335-3e13-44c6-844f-318168230b30","Type":"ContainerStarted","Data":"e40e55c25a4094af294abe198a08c4afe5a2bd3f4829dc3c3a70182d263102d4"} Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.862635 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.862848 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.362817803 +0000 UTC m=+142.412609114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.863045 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.863118 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-utilities\") pod \"certified-operators-7cqfs\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.863219 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mlsq\" (UniqueName: \"kubernetes.io/projected/d53216ca-283e-4346-ace1-9fab11518bfd-kube-api-access-7mlsq\") pod \"certified-operators-7cqfs\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.863262 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-catalog-content\") pod \"certified-operators-7cqfs\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.863376 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.363343217 +0000 UTC m=+142.413134538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.863953 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-catalog-content\") pod \"certified-operators-7cqfs\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.864310 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-utilities\") pod \"certified-operators-7cqfs\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.888907 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mlsq\" (UniqueName: \"kubernetes.io/projected/d53216ca-283e-4346-ace1-9fab11518bfd-kube-api-access-7mlsq\") pod \"certified-operators-7cqfs\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.964284 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.964471 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.464443815 +0000 UTC m=+142.514235126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.966692 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:34 crc kubenswrapper[4805]: E1128 15:28:34.968237 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.468216353 +0000 UTC m=+142.518007694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:34 crc kubenswrapper[4805]: I1128 15:28:34.975932 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.023638 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:35 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:35 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:35 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.023704 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.069008 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.069150 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.569124177 +0000 UTC m=+142.618915478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.069311 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.069662 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.569654851 +0000 UTC m=+142.619446162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.174461 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.174605 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.674588199 +0000 UTC m=+142.724379510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.174714 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.175018 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.675010351 +0000 UTC m=+142.724801662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.283415 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.283633 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.783594103 +0000 UTC m=+142.833385454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.284158 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.284608 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.784590101 +0000 UTC m=+142.834381412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.330233 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-catalog-content\") pod \"certified-operators-dhzqc\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.334589 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gssxt\" (UniqueName: \"kubernetes.io/projected/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-kube-api-access-gssxt\") pod \"certified-operators-dhzqc\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.385752 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.386945 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.886928115 +0000 UTC m=+142.936719426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.392721 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.404487 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.491860 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ddd7ac2-635b-4fba-9765-a81039204b8f-secret-volume\") pod \"4ddd7ac2-635b-4fba-9765-a81039204b8f\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.492063 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xpkz\" (UniqueName: \"kubernetes.io/projected/4ddd7ac2-635b-4fba-9765-a81039204b8f-kube-api-access-7xpkz\") pod \"4ddd7ac2-635b-4fba-9765-a81039204b8f\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.492114 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ddd7ac2-635b-4fba-9765-a81039204b8f-config-volume\") pod \"4ddd7ac2-635b-4fba-9765-a81039204b8f\" (UID: \"4ddd7ac2-635b-4fba-9765-a81039204b8f\") " Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.492396 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.493386 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ddd7ac2-635b-4fba-9765-a81039204b8f-config-volume" (OuterVolumeSpecName: "config-volume") pod "4ddd7ac2-635b-4fba-9765-a81039204b8f" (UID: "4ddd7ac2-635b-4fba-9765-a81039204b8f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.494629 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:35.994607591 +0000 UTC m=+143.044398982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.503581 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ddd7ac2-635b-4fba-9765-a81039204b8f-kube-api-access-7xpkz" (OuterVolumeSpecName: "kube-api-access-7xpkz") pod "4ddd7ac2-635b-4fba-9765-a81039204b8f" (UID: "4ddd7ac2-635b-4fba-9765-a81039204b8f"). InnerVolumeSpecName "kube-api-access-7xpkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.503665 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ddd7ac2-635b-4fba-9765-a81039204b8f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4ddd7ac2-635b-4fba-9765-a81039204b8f" (UID: "4ddd7ac2-635b-4fba-9765-a81039204b8f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.521932 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.594260 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.594972 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ddd7ac2-635b-4fba-9765-a81039204b8f-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.594989 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ddd7ac2-635b-4fba-9765-a81039204b8f-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.595002 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xpkz\" (UniqueName: \"kubernetes.io/projected/4ddd7ac2-635b-4fba-9765-a81039204b8f-kube-api-access-7xpkz\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.595081 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.095063581 +0000 UTC m=+143.144854892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.696231 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.696582 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.1965709 +0000 UTC m=+143.246362211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.726893 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rgs2f"] Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.798034 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.798387 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.298372649 +0000 UTC m=+143.348163960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.802605 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dhzqc"] Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.877564 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhzqc" event={"ID":"bf218599-d9d7-4e0e-a2ab-0949ded9fa58","Type":"ContainerStarted","Data":"71cc22f73e7b72ee7fcc570e09b6438c0cfdab3975a9f0d0268b6d4731971286"} Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.900189 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:35 crc kubenswrapper[4805]: E1128 15:28:35.900553 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.400539647 +0000 UTC m=+143.450330958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.909826 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" event={"ID":"4ddd7ac2-635b-4fba-9765-a81039204b8f","Type":"ContainerDied","Data":"ab82356be96a522e8277bac4728b5a1aea4711adfa485fcd0569a8ed2618c607"} Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.909860 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab82356be96a522e8277bac4728b5a1aea4711adfa485fcd0569a8ed2618c607" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.909945 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc" Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.928806 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgs2f" event={"ID":"c6b04434-a30b-4e22-8c5a-9cdaca124466","Type":"ContainerStarted","Data":"00319e1ad1bf63777319714659b1b6f01e340948d65aec8acb8a3db4d3be68a2"} Nov 28 15:28:35 crc kubenswrapper[4805]: I1128 15:28:35.991898 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7cqfs"] Nov 28 15:28:36 crc kubenswrapper[4805]: W1128 15:28:36.000449 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd53216ca_283e_4346_ace1_9fab11518bfd.slice/crio-9f59f8b0e43942fa7d88be3d07651c9215957a017d73a3bfa1a4bcd31a5373e7 WatchSource:0}: Error finding container 9f59f8b0e43942fa7d88be3d07651c9215957a017d73a3bfa1a4bcd31a5373e7: Status 404 returned error can't find the container with id 9f59f8b0e43942fa7d88be3d07651c9215957a017d73a3bfa1a4bcd31a5373e7 Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.000907 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.001242 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.501192283 +0000 UTC m=+143.550983594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.001412 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.001828 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.50181573 +0000 UTC m=+143.551607041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.029796 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:36 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:36 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:36 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.030186 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.102866 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.103045 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.603019402 +0000 UTC m=+143.652810723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.103201 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.103581 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.603567978 +0000 UTC m=+143.653359309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.204390 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.204653 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.704624475 +0000 UTC m=+143.754415806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.261247 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s2fzf"] Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.261750 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddd7ac2-635b-4fba-9765-a81039204b8f" containerName="collect-profiles" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.261827 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddd7ac2-635b-4fba-9765-a81039204b8f" containerName="collect-profiles" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.261975 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ddd7ac2-635b-4fba-9765-a81039204b8f" containerName="collect-profiles" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.262918 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.268248 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.270266 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.271678 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.275956 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.280623 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2fzf"] Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.282784 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.285932 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.305998 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.306311 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.806295469 +0000 UTC m=+143.856086780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.407062 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.407278 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:36.907247893 +0000 UTC m=+143.957039214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.407337 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68a4ee41-f904-4013-a9d2-47ad12487870-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"68a4ee41-f904-4013-a9d2-47ad12487870\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.407472 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-catalog-content\") pod \"redhat-marketplace-s2fzf\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.407553 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlkbz\" (UniqueName: \"kubernetes.io/projected/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-kube-api-access-wlkbz\") pod \"redhat-marketplace-s2fzf\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.407783 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68a4ee41-f904-4013-a9d2-47ad12487870-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"68a4ee41-f904-4013-a9d2-47ad12487870\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.407834 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-utilities\") pod \"redhat-marketplace-s2fzf\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.497062 4805 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.508964 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-catalog-content\") pod \"redhat-marketplace-s2fzf\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.509032 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlkbz\" (UniqueName: \"kubernetes.io/projected/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-kube-api-access-wlkbz\") pod \"redhat-marketplace-s2fzf\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.509130 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68a4ee41-f904-4013-a9d2-47ad12487870-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"68a4ee41-f904-4013-a9d2-47ad12487870\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.509157 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-utilities\") pod \"redhat-marketplace-s2fzf\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.509226 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.509262 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68a4ee41-f904-4013-a9d2-47ad12487870-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"68a4ee41-f904-4013-a9d2-47ad12487870\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.510334 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-catalog-content\") pod \"redhat-marketplace-s2fzf\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.510650 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-utilities\") pod \"redhat-marketplace-s2fzf\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.510664 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.010648987 +0000 UTC m=+144.060440298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.510699 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68a4ee41-f904-4013-a9d2-47ad12487870-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"68a4ee41-f904-4013-a9d2-47ad12487870\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.534026 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlkbz\" (UniqueName: \"kubernetes.io/projected/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-kube-api-access-wlkbz\") pod \"redhat-marketplace-s2fzf\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.534513 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68a4ee41-f904-4013-a9d2-47ad12487870-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"68a4ee41-f904-4013-a9d2-47ad12487870\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.577025 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.590620 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.609834 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.610149 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.110135729 +0000 UTC m=+144.159927030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.659730 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tgr5n"] Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.661438 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.669237 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgr5n"] Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.711794 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.712541 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.212527554 +0000 UTC m=+144.262318865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.800375 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2fzf"] Nov 28 15:28:36 crc kubenswrapper[4805]: W1128 15:28:36.803486 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeff8293e_510a_4cf5_afe2_5fefc4da0d3f.slice/crio-3afba656b135a59d75e8f19b1543a8b4738661d9bc002869e66af7a4c486779b WatchSource:0}: Error finding container 3afba656b135a59d75e8f19b1543a8b4738661d9bc002869e66af7a4c486779b: Status 404 returned error can't find the container with id 3afba656b135a59d75e8f19b1543a8b4738661d9bc002869e66af7a4c486779b Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.813244 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.813574 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l78sb\" (UniqueName: \"kubernetes.io/projected/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-kube-api-access-l78sb\") pod \"redhat-marketplace-tgr5n\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.813623 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-catalog-content\") pod \"redhat-marketplace-tgr5n\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.813716 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-utilities\") pod \"redhat-marketplace-tgr5n\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.813812 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.313797257 +0000 UTC m=+144.363588568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.828844 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.828886 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.836422 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.839346 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:36 crc kubenswrapper[4805]: W1128 15:28:36.847089 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod68a4ee41_f904_4013_a9d2_47ad12487870.slice/crio-b15098a5f5d066194b157162f6a6ae8c321be65b2b2f4573e278dd0fa57e4774 WatchSource:0}: Error finding container b15098a5f5d066194b157162f6a6ae8c321be65b2b2f4573e278dd0fa57e4774: Status 404 returned error can't find the container with id b15098a5f5d066194b157162f6a6ae8c321be65b2b2f4573e278dd0fa57e4774 Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.874769 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.874811 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.880755 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.880797 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.881747 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.882136 4805 patch_prober.go:28] interesting pod/console-f9d7485db-7bprj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.882199 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7bprj" podUID="69b46aab-afdf-4551-b0b4-b8d4f8f97831" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.890289 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-2b2hq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.890378 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2b2hq" podUID="97ea3cfc-6d2a-417f-a397-34a5760d0a9e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.890474 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-2b2hq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.890519 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2b2hq" podUID="97ea3cfc-6d2a-417f-a397-34a5760d0a9e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.915136 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.915211 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-utilities\") pod \"redhat-marketplace-tgr5n\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.915263 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l78sb\" (UniqueName: \"kubernetes.io/projected/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-kube-api-access-l78sb\") pod \"redhat-marketplace-tgr5n\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.915313 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-catalog-content\") pod \"redhat-marketplace-tgr5n\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: E1128 15:28:36.915544 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.415527253 +0000 UTC m=+144.465318564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.916032 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-utilities\") pod \"redhat-marketplace-tgr5n\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.916097 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-catalog-content\") pod \"redhat-marketplace-tgr5n\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.936982 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l78sb\" (UniqueName: \"kubernetes.io/projected/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-kube-api-access-l78sb\") pod \"redhat-marketplace-tgr5n\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.938760 4805 generic.go:334] "Generic (PLEG): container finished" podID="db028335-3e13-44c6-844f-318168230b30" containerID="5bb049dff1328fb61ad59f6d28dd769a3323dc3facceb6099f58a0130398da4e" exitCode=0 Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.939008 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d4l4" event={"ID":"db028335-3e13-44c6-844f-318168230b30","Type":"ContainerDied","Data":"5bb049dff1328fb61ad59f6d28dd769a3323dc3facceb6099f58a0130398da4e"} Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.940287 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.947219 4805 generic.go:334] "Generic (PLEG): container finished" podID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerID="b2d20787fca4d3655496a5b84092c393da251ee288de74552d41e4d96e404918" exitCode=0 Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.947293 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgs2f" event={"ID":"c6b04434-a30b-4e22-8c5a-9cdaca124466","Type":"ContainerDied","Data":"b2d20787fca4d3655496a5b84092c393da251ee288de74552d41e4d96e404918"} Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.957954 4805 generic.go:334] "Generic (PLEG): container finished" podID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerID="8b142cd70a956ee827eb8f17fa59358d1f273ab2ef0f7d0adecded046621aae4" exitCode=0 Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.958088 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhzqc" event={"ID":"bf218599-d9d7-4e0e-a2ab-0949ded9fa58","Type":"ContainerDied","Data":"8b142cd70a956ee827eb8f17fa59358d1f273ab2ef0f7d0adecded046621aae4"} Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.966652 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"68a4ee41-f904-4013-a9d2-47ad12487870","Type":"ContainerStarted","Data":"b15098a5f5d066194b157162f6a6ae8c321be65b2b2f4573e278dd0fa57e4774"} Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.969957 4805 generic.go:334] "Generic (PLEG): container finished" podID="d53216ca-283e-4346-ace1-9fab11518bfd" containerID="d7c90d015f58ac95ab50438f250cd78877ac98fe6b856cf8362686840dd945fe" exitCode=0 Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.970090 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cqfs" event={"ID":"d53216ca-283e-4346-ace1-9fab11518bfd","Type":"ContainerDied","Data":"d7c90d015f58ac95ab50438f250cd78877ac98fe6b856cf8362686840dd945fe"} Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.970228 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cqfs" event={"ID":"d53216ca-283e-4346-ace1-9fab11518bfd","Type":"ContainerStarted","Data":"9f59f8b0e43942fa7d88be3d07651c9215957a017d73a3bfa1a4bcd31a5373e7"} Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.973884 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" event={"ID":"e2692e90-917d-4467-825d-5e277ce862bc","Type":"ContainerStarted","Data":"ce92b44c73a16fb7630a0d112c0b240bdcab2f548b689d8ede658d8bb09773c8"} Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.974653 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" event={"ID":"e2692e90-917d-4467-825d-5e277ce862bc","Type":"ContainerStarted","Data":"27e4b3ddddfc155201581f89313056d471a1f84bf65eaf446e21364d51f2b6a3"} Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.984235 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2fzf" event={"ID":"eff8293e-510a-4cf5-afe2-5fefc4da0d3f","Type":"ContainerStarted","Data":"3afba656b135a59d75e8f19b1543a8b4738661d9bc002869e66af7a4c486779b"} Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.988700 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-scnxw" Nov 28 15:28:36 crc kubenswrapper[4805]: I1128 15:28:36.990451 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpvb4" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.016256 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:37 crc kubenswrapper[4805]: E1128 15:28:37.016411 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.516386625 +0000 UTC m=+144.566177936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.019101 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.023395 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:37 crc kubenswrapper[4805]: E1128 15:28:37.023553 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.523533458 +0000 UTC m=+144.573324799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.025109 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:37 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:37 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:37 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.025319 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.030956 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.126160 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:37 crc kubenswrapper[4805]: E1128 15:28:37.136316 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.636282249 +0000 UTC m=+144.686073550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.227913 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:37 crc kubenswrapper[4805]: E1128 15:28:37.228443 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.728425511 +0000 UTC m=+144.778216822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5vvq" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.277760 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s6fmm"] Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.279153 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.281521 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s6fmm"] Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.281758 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.328650 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:37 crc kubenswrapper[4805]: E1128 15:28:37.329181 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 15:28:37.829152569 +0000 UTC m=+144.878943870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.365477 4805 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-28T15:28:36.497092319Z","Handler":null,"Name":""} Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.396728 4805 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.397259 4805 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.430094 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lcm5\" (UniqueName: \"kubernetes.io/projected/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-kube-api-access-9lcm5\") pod \"redhat-operators-s6fmm\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.430167 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-utilities\") pod \"redhat-operators-s6fmm\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.430192 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-catalog-content\") pod \"redhat-operators-s6fmm\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.430282 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.434930 4805 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.434980 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.496151 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5vvq\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.524335 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgr5n"] Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.532476 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.532730 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-utilities\") pod \"redhat-operators-s6fmm\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.532762 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-catalog-content\") pod \"redhat-operators-s6fmm\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.532840 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lcm5\" (UniqueName: \"kubernetes.io/projected/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-kube-api-access-9lcm5\") pod \"redhat-operators-s6fmm\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.533685 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-utilities\") pod \"redhat-operators-s6fmm\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.533866 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-catalog-content\") pod \"redhat-operators-s6fmm\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.605696 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lcm5\" (UniqueName: \"kubernetes.io/projected/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-kube-api-access-9lcm5\") pod \"redhat-operators-s6fmm\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.613799 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.614004 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.700762 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hm6n6"] Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.701837 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.722598 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hm6n6"] Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.773648 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.842802 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxpjb\" (UniqueName: \"kubernetes.io/projected/e9089841-40aa-479f-b56e-421629f24685-kube-api-access-sxpjb\") pod \"redhat-operators-hm6n6\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.842915 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-catalog-content\") pod \"redhat-operators-hm6n6\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.842953 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-utilities\") pod \"redhat-operators-hm6n6\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.947169 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-catalog-content\") pod \"redhat-operators-hm6n6\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.947540 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-utilities\") pod \"redhat-operators-hm6n6\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.947619 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxpjb\" (UniqueName: \"kubernetes.io/projected/e9089841-40aa-479f-b56e-421629f24685-kube-api-access-sxpjb\") pod \"redhat-operators-hm6n6\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.948341 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-catalog-content\") pod \"redhat-operators-hm6n6\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.948571 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-utilities\") pod \"redhat-operators-hm6n6\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:37 crc kubenswrapper[4805]: I1128 15:28:37.984559 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxpjb\" (UniqueName: \"kubernetes.io/projected/e9089841-40aa-479f-b56e-421629f24685-kube-api-access-sxpjb\") pod \"redhat-operators-hm6n6\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.023830 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:38 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:38 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:38 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.023886 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.029239 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.035816 4805 generic.go:334] "Generic (PLEG): container finished" podID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerID="9d477a9f7cf06a01e0bda14061263382874f8c57d67d0beeeeb44f4b6a1d4e60" exitCode=0 Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.036138 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgr5n" event={"ID":"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa","Type":"ContainerDied","Data":"9d477a9f7cf06a01e0bda14061263382874f8c57d67d0beeeeb44f4b6a1d4e60"} Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.036206 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgr5n" event={"ID":"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa","Type":"ContainerStarted","Data":"9f4b512043c98bcced425657cb2268c5affabd65f19d2539a3d3c0e24b8b841d"} Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.059590 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"68a4ee41-f904-4013-a9d2-47ad12487870","Type":"ContainerStarted","Data":"b38f21edf963e45e70b7f8811651e342b8f99e60f61897acf752bd430f9eab57"} Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.071594 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" event={"ID":"e2692e90-917d-4467-825d-5e277ce862bc","Type":"ContainerStarted","Data":"f678b0757c7f885364e2a20b8e15c89360149fa5b957cbf69473fcff0e6bf048"} Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.075700 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.075685016 podStartE2EDuration="2.075685016s" podCreationTimestamp="2025-11-28 15:28:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:38.073481943 +0000 UTC m=+145.123273254" watchObservedRunningTime="2025-11-28 15:28:38.075685016 +0000 UTC m=+145.125476327" Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.077061 4805 generic.go:334] "Generic (PLEG): container finished" podID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerID="aa867351a96ffb824c7c8c4ee504065aeecb3924e0437e115806297ce9d380fa" exitCode=0 Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.077808 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2fzf" event={"ID":"eff8293e-510a-4cf5-afe2-5fefc4da0d3f","Type":"ContainerDied","Data":"aa867351a96ffb824c7c8c4ee504065aeecb3924e0437e115806297ce9d380fa"} Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.121099 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-t4l7k" podStartSLOduration=14.121080543 podStartE2EDuration="14.121080543s" podCreationTimestamp="2025-11-28 15:28:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:38.095099731 +0000 UTC m=+145.144891042" watchObservedRunningTime="2025-11-28 15:28:38.121080543 +0000 UTC m=+145.170871854" Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.198652 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s6fmm"] Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.310150 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.327851 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pqxhf" Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.330872 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5vvq"] Nov 28 15:28:38 crc kubenswrapper[4805]: I1128 15:28:38.361601 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hm6n6"] Nov 28 15:28:38 crc kubenswrapper[4805]: W1128 15:28:38.397250 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0edd21b_3a3d_41c4_8e8b_af747bd69575.slice/crio-48be2fe808df3375cd359af318b0f4d9eda885373e34728ac03484f17b7cb1d5 WatchSource:0}: Error finding container 48be2fe808df3375cd359af318b0f4d9eda885373e34728ac03484f17b7cb1d5: Status 404 returned error can't find the container with id 48be2fe808df3375cd359af318b0f4d9eda885373e34728ac03484f17b7cb1d5 Nov 28 15:28:38 crc kubenswrapper[4805]: W1128 15:28:38.521766 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9089841_40aa_479f_b56e_421629f24685.slice/crio-ffdf9858dcc79eff07c7aa05400f1a16d6e641ec0b13ff4a08b3bdc1149fcf6d WatchSource:0}: Error finding container ffdf9858dcc79eff07c7aa05400f1a16d6e641ec0b13ff4a08b3bdc1149fcf6d: Status 404 returned error can't find the container with id ffdf9858dcc79eff07c7aa05400f1a16d6e641ec0b13ff4a08b3bdc1149fcf6d Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.024688 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:39 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:39 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:39 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.024795 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.091418 4805 generic.go:334] "Generic (PLEG): container finished" podID="68a4ee41-f904-4013-a9d2-47ad12487870" containerID="b38f21edf963e45e70b7f8811651e342b8f99e60f61897acf752bd430f9eab57" exitCode=0 Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.091488 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"68a4ee41-f904-4013-a9d2-47ad12487870","Type":"ContainerDied","Data":"b38f21edf963e45e70b7f8811651e342b8f99e60f61897acf752bd430f9eab57"} Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.095890 4805 generic.go:334] "Generic (PLEG): container finished" podID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerID="47d629bf37f80023c3e1c47273a4700feac831b3d8c5c56216e89bb4a67a0406" exitCode=0 Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.095950 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6fmm" event={"ID":"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07","Type":"ContainerDied","Data":"47d629bf37f80023c3e1c47273a4700feac831b3d8c5c56216e89bb4a67a0406"} Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.096121 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6fmm" event={"ID":"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07","Type":"ContainerStarted","Data":"5a827dd1ed486794b97b8956f8e2aeaf45cf10b6a237e6d2dfd792bd559ace05"} Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.101072 4805 generic.go:334] "Generic (PLEG): container finished" podID="e9089841-40aa-479f-b56e-421629f24685" containerID="cbaa3ce5d78508340baff32d7fffeab7879c9eae9ca3ef66f4a9da550fa9a8c2" exitCode=0 Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.101162 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm6n6" event={"ID":"e9089841-40aa-479f-b56e-421629f24685","Type":"ContainerDied","Data":"cbaa3ce5d78508340baff32d7fffeab7879c9eae9ca3ef66f4a9da550fa9a8c2"} Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.101185 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm6n6" event={"ID":"e9089841-40aa-479f-b56e-421629f24685","Type":"ContainerStarted","Data":"ffdf9858dcc79eff07c7aa05400f1a16d6e641ec0b13ff4a08b3bdc1149fcf6d"} Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.105409 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" event={"ID":"c0edd21b-3a3d-41c4-8e8b-af747bd69575","Type":"ContainerStarted","Data":"48be2fe808df3375cd359af318b0f4d9eda885373e34728ac03484f17b7cb1d5"} Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.220645 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.281128 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.281191 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.281255 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.281322 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.282201 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.306174 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.307304 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.311552 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.317628 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.336936 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:39 crc kubenswrapper[4805]: I1128 15:28:39.355712 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.023524 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:40 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:40 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:40 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.023941 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:40 crc kubenswrapper[4805]: W1128 15:28:40.092592 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-23329ce38e91afc8010cccd6a39cd302c426f9e118e7052f1275abb754382f6c WatchSource:0}: Error finding container 23329ce38e91afc8010cccd6a39cd302c426f9e118e7052f1275abb754382f6c: Status 404 returned error can't find the container with id 23329ce38e91afc8010cccd6a39cd302c426f9e118e7052f1275abb754382f6c Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.155434 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"23329ce38e91afc8010cccd6a39cd302c426f9e118e7052f1275abb754382f6c"} Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.158391 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" event={"ID":"c0edd21b-3a3d-41c4-8e8b-af747bd69575","Type":"ContainerStarted","Data":"85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878"} Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.162393 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7e7cb659af276a61d5262a2707ee570480a80ca95518f1470b41b9ec179e8db4"} Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.173671 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"730c171024b30815740a1a44b017afcbb68dd0817c39b933f40c2d64e8f939ed"} Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.195877 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" podStartSLOduration=128.195849754 podStartE2EDuration="2m8.195849754s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:28:40.19190086 +0000 UTC m=+147.241692171" watchObservedRunningTime="2025-11-28 15:28:40.195849754 +0000 UTC m=+147.245641055" Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.332185 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-bpmbs" Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.493590 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.605007 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68a4ee41-f904-4013-a9d2-47ad12487870-kubelet-dir\") pod \"68a4ee41-f904-4013-a9d2-47ad12487870\" (UID: \"68a4ee41-f904-4013-a9d2-47ad12487870\") " Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.605186 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68a4ee41-f904-4013-a9d2-47ad12487870-kube-api-access\") pod \"68a4ee41-f904-4013-a9d2-47ad12487870\" (UID: \"68a4ee41-f904-4013-a9d2-47ad12487870\") " Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.605166 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68a4ee41-f904-4013-a9d2-47ad12487870-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "68a4ee41-f904-4013-a9d2-47ad12487870" (UID: "68a4ee41-f904-4013-a9d2-47ad12487870"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.613729 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a4ee41-f904-4013-a9d2-47ad12487870-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "68a4ee41-f904-4013-a9d2-47ad12487870" (UID: "68a4ee41-f904-4013-a9d2-47ad12487870"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.710465 4805 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68a4ee41-f904-4013-a9d2-47ad12487870-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:40 crc kubenswrapper[4805]: I1128 15:28:40.710515 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68a4ee41-f904-4013-a9d2-47ad12487870-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:41 crc kubenswrapper[4805]: I1128 15:28:41.026568 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:41 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:41 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:41 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:41 crc kubenswrapper[4805]: I1128 15:28:41.026784 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:41 crc kubenswrapper[4805]: I1128 15:28:41.060625 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:28:41 crc kubenswrapper[4805]: I1128 15:28:41.060710 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:28:41 crc kubenswrapper[4805]: I1128 15:28:41.189349 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"68a4ee41-f904-4013-a9d2-47ad12487870","Type":"ContainerDied","Data":"b15098a5f5d066194b157162f6a6ae8c321be65b2b2f4573e278dd0fa57e4774"} Nov 28 15:28:41 crc kubenswrapper[4805]: I1128 15:28:41.189415 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b15098a5f5d066194b157162f6a6ae8c321be65b2b2f4573e278dd0fa57e4774" Nov 28 15:28:41 crc kubenswrapper[4805]: I1128 15:28:41.189489 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 15:28:41 crc kubenswrapper[4805]: I1128 15:28:41.195586 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c9bc719e5eb746f8f4a0e12e88428b01bf4b44a0a2544d46ccc254257f833b91"} Nov 28 15:28:41 crc kubenswrapper[4805]: I1128 15:28:41.195781 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:28:42 crc kubenswrapper[4805]: I1128 15:28:42.022752 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:42 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:42 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:42 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:42 crc kubenswrapper[4805]: I1128 15:28:42.023328 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:42 crc kubenswrapper[4805]: I1128 15:28:42.208685 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e6e839d27f5f830e5393b9327238d9a1e81e26a7ababe9fd6aac205a58a9104e"} Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.024534 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:43 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:43 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:43 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.024616 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.226720 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"414652158f3d49a3e75b65b44136cfd9f2aaed7cc47bf801d030668ecb200c84"} Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.227520 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.385924 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 15:28:43 crc kubenswrapper[4805]: E1128 15:28:43.386214 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a4ee41-f904-4013-a9d2-47ad12487870" containerName="pruner" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.386231 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a4ee41-f904-4013-a9d2-47ad12487870" containerName="pruner" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.386353 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a4ee41-f904-4013-a9d2-47ad12487870" containerName="pruner" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.386813 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.391880 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.395913 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.401022 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.473802 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4e71df1-307a-4e5b-96fd-25e64108c715-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e4e71df1-307a-4e5b-96fd-25e64108c715\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.473879 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e4e71df1-307a-4e5b-96fd-25e64108c715-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e4e71df1-307a-4e5b-96fd-25e64108c715\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.575336 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4e71df1-307a-4e5b-96fd-25e64108c715-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e4e71df1-307a-4e5b-96fd-25e64108c715\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.575416 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e4e71df1-307a-4e5b-96fd-25e64108c715-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e4e71df1-307a-4e5b-96fd-25e64108c715\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.575564 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e4e71df1-307a-4e5b-96fd-25e64108c715-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e4e71df1-307a-4e5b-96fd-25e64108c715\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.601333 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4e71df1-307a-4e5b-96fd-25e64108c715-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e4e71df1-307a-4e5b-96fd-25e64108c715\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:43 crc kubenswrapper[4805]: I1128 15:28:43.712130 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:44 crc kubenswrapper[4805]: I1128 15:28:44.023559 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:44 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:44 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:44 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:44 crc kubenswrapper[4805]: I1128 15:28:44.023628 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:44 crc kubenswrapper[4805]: I1128 15:28:44.241718 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 15:28:44 crc kubenswrapper[4805]: W1128 15:28:44.270538 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode4e71df1_307a_4e5b_96fd_25e64108c715.slice/crio-db9ec17563be5b8fab967f8c53ad02d0b3d51abdc640ce290c22d26c95bda46c WatchSource:0}: Error finding container db9ec17563be5b8fab967f8c53ad02d0b3d51abdc640ce290c22d26c95bda46c: Status 404 returned error can't find the container with id db9ec17563be5b8fab967f8c53ad02d0b3d51abdc640ce290c22d26c95bda46c Nov 28 15:28:45 crc kubenswrapper[4805]: I1128 15:28:45.024797 4805 patch_prober.go:28] interesting pod/router-default-5444994796-lfq9l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 15:28:45 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Nov 28 15:28:45 crc kubenswrapper[4805]: [+]process-running ok Nov 28 15:28:45 crc kubenswrapper[4805]: healthz check failed Nov 28 15:28:45 crc kubenswrapper[4805]: I1128 15:28:45.025485 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfq9l" podUID="e6d68711-24e1-46d1-95bb-31974c24b299" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 15:28:45 crc kubenswrapper[4805]: I1128 15:28:45.249036 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e4e71df1-307a-4e5b-96fd-25e64108c715","Type":"ContainerStarted","Data":"db9ec17563be5b8fab967f8c53ad02d0b3d51abdc640ce290c22d26c95bda46c"} Nov 28 15:28:46 crc kubenswrapper[4805]: I1128 15:28:46.024244 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:46 crc kubenswrapper[4805]: I1128 15:28:46.030777 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lfq9l" Nov 28 15:28:46 crc kubenswrapper[4805]: I1128 15:28:46.260189 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e4e71df1-307a-4e5b-96fd-25e64108c715","Type":"ContainerStarted","Data":"8a168d23d734ca3980788ae634bd8f76a80490592b0f2b4e6a89ad8b6174be3f"} Nov 28 15:28:46 crc kubenswrapper[4805]: I1128 15:28:46.880998 4805 patch_prober.go:28] interesting pod/console-f9d7485db-7bprj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 28 15:28:46 crc kubenswrapper[4805]: I1128 15:28:46.881312 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7bprj" podUID="69b46aab-afdf-4551-b0b4-b8d4f8f97831" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 28 15:28:46 crc kubenswrapper[4805]: I1128 15:28:46.901448 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-2b2hq" Nov 28 15:28:47 crc kubenswrapper[4805]: I1128 15:28:47.298266 4805 generic.go:334] "Generic (PLEG): container finished" podID="e4e71df1-307a-4e5b-96fd-25e64108c715" containerID="8a168d23d734ca3980788ae634bd8f76a80490592b0f2b4e6a89ad8b6174be3f" exitCode=0 Nov 28 15:28:47 crc kubenswrapper[4805]: I1128 15:28:47.298344 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e4e71df1-307a-4e5b-96fd-25e64108c715","Type":"ContainerDied","Data":"8a168d23d734ca3980788ae634bd8f76a80490592b0f2b4e6a89ad8b6174be3f"} Nov 28 15:28:54 crc kubenswrapper[4805]: I1128 15:28:54.560068 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:54 crc kubenswrapper[4805]: I1128 15:28:54.566487 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18cf186f-76f6-47bc-8db9-c9b7be3aaf09-metrics-certs\") pod \"network-metrics-daemon-fplc8\" (UID: \"18cf186f-76f6-47bc-8db9-c9b7be3aaf09\") " pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:54 crc kubenswrapper[4805]: I1128 15:28:54.625610 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fplc8" Nov 28 15:28:56 crc kubenswrapper[4805]: I1128 15:28:56.662604 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:56 crc kubenswrapper[4805]: I1128 15:28:56.813888 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e4e71df1-307a-4e5b-96fd-25e64108c715-kubelet-dir\") pod \"e4e71df1-307a-4e5b-96fd-25e64108c715\" (UID: \"e4e71df1-307a-4e5b-96fd-25e64108c715\") " Nov 28 15:28:56 crc kubenswrapper[4805]: I1128 15:28:56.814053 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4e71df1-307a-4e5b-96fd-25e64108c715-kube-api-access\") pod \"e4e71df1-307a-4e5b-96fd-25e64108c715\" (UID: \"e4e71df1-307a-4e5b-96fd-25e64108c715\") " Nov 28 15:28:56 crc kubenswrapper[4805]: I1128 15:28:56.814078 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4e71df1-307a-4e5b-96fd-25e64108c715-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e4e71df1-307a-4e5b-96fd-25e64108c715" (UID: "e4e71df1-307a-4e5b-96fd-25e64108c715"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:28:56 crc kubenswrapper[4805]: I1128 15:28:56.814311 4805 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e4e71df1-307a-4e5b-96fd-25e64108c715-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:56 crc kubenswrapper[4805]: I1128 15:28:56.817946 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4e71df1-307a-4e5b-96fd-25e64108c715-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e4e71df1-307a-4e5b-96fd-25e64108c715" (UID: "e4e71df1-307a-4e5b-96fd-25e64108c715"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:28:56 crc kubenswrapper[4805]: I1128 15:28:56.884401 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:56 crc kubenswrapper[4805]: I1128 15:28:56.888582 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:28:56 crc kubenswrapper[4805]: I1128 15:28:56.914989 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4e71df1-307a-4e5b-96fd-25e64108c715-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:57 crc kubenswrapper[4805]: I1128 15:28:57.383695 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e4e71df1-307a-4e5b-96fd-25e64108c715","Type":"ContainerDied","Data":"db9ec17563be5b8fab967f8c53ad02d0b3d51abdc640ce290c22d26c95bda46c"} Nov 28 15:28:57 crc kubenswrapper[4805]: I1128 15:28:57.383742 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db9ec17563be5b8fab967f8c53ad02d0b3d51abdc640ce290c22d26c95bda46c" Nov 28 15:28:57 crc kubenswrapper[4805]: I1128 15:28:57.383713 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 15:28:57 crc kubenswrapper[4805]: I1128 15:28:57.780210 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:29:07 crc kubenswrapper[4805]: I1128 15:29:07.097413 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8zrtm" Nov 28 15:29:11 crc kubenswrapper[4805]: I1128 15:29:11.059949 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:29:11 crc kubenswrapper[4805]: I1128 15:29:11.060208 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.423822 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 15:29:18 crc kubenswrapper[4805]: E1128 15:29:18.424806 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e71df1-307a-4e5b-96fd-25e64108c715" containerName="pruner" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.424826 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e71df1-307a-4e5b-96fd-25e64108c715" containerName="pruner" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.424992 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4e71df1-307a-4e5b-96fd-25e64108c715" containerName="pruner" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.425491 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.434941 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.435078 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.443970 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.530728 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7779b1cc-4309-4cc8-b634-903dec64f766-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7779b1cc-4309-4cc8-b634-903dec64f766\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.530899 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7779b1cc-4309-4cc8-b634-903dec64f766-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7779b1cc-4309-4cc8-b634-903dec64f766\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.632160 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7779b1cc-4309-4cc8-b634-903dec64f766-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7779b1cc-4309-4cc8-b634-903dec64f766\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.632235 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7779b1cc-4309-4cc8-b634-903dec64f766-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7779b1cc-4309-4cc8-b634-903dec64f766\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.632283 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7779b1cc-4309-4cc8-b634-903dec64f766-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7779b1cc-4309-4cc8-b634-903dec64f766\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.664642 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7779b1cc-4309-4cc8-b634-903dec64f766-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7779b1cc-4309-4cc8-b634-903dec64f766\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:18 crc kubenswrapper[4805]: I1128 15:29:18.761161 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:19 crc kubenswrapper[4805]: I1128 15:29:19.484531 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 15:29:23 crc kubenswrapper[4805]: I1128 15:29:23.967979 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 15:29:23 crc kubenswrapper[4805]: I1128 15:29:23.969850 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:23 crc kubenswrapper[4805]: I1128 15:29:23.976937 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.103797 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72212593-ffb8-48a1-a673-5532833d9134-kube-api-access\") pod \"installer-9-crc\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.104340 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-kubelet-dir\") pod \"installer-9-crc\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.104421 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-var-lock\") pod \"installer-9-crc\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.208815 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-kubelet-dir\") pod \"installer-9-crc\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.208901 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-var-lock\") pod \"installer-9-crc\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.208951 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72212593-ffb8-48a1-a673-5532833d9134-kube-api-access\") pod \"installer-9-crc\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.208981 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-kubelet-dir\") pod \"installer-9-crc\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.209044 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-var-lock\") pod \"installer-9-crc\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.246712 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72212593-ffb8-48a1-a673-5532833d9134-kube-api-access\") pod \"installer-9-crc\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:24 crc kubenswrapper[4805]: I1128 15:29:24.300608 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:29:26 crc kubenswrapper[4805]: E1128 15:29:26.713185 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 15:29:26 crc kubenswrapper[4805]: E1128 15:29:26.713430 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lcm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-s6fmm_openshift-marketplace(bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:29:26 crc kubenswrapper[4805]: E1128 15:29:26.714613 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-s6fmm" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" Nov 28 15:29:29 crc kubenswrapper[4805]: E1128 15:29:29.920866 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-s6fmm" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" Nov 28 15:29:40 crc kubenswrapper[4805]: E1128 15:29:40.725585 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 15:29:40 crc kubenswrapper[4805]: E1128 15:29:40.727240 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l78sb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-tgr5n_openshift-marketplace(7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:29:40 crc kubenswrapper[4805]: E1128 15:29:40.728531 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-tgr5n" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" Nov 28 15:29:40 crc kubenswrapper[4805]: E1128 15:29:40.946776 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 15:29:40 crc kubenswrapper[4805]: E1128 15:29:40.947215 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sxpjb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-hm6n6_openshift-marketplace(e9089841-40aa-479f-b56e-421629f24685): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:29:40 crc kubenswrapper[4805]: E1128 15:29:40.948452 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-hm6n6" podUID="e9089841-40aa-479f-b56e-421629f24685" Nov 28 15:29:41 crc kubenswrapper[4805]: I1128 15:29:41.060545 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:29:41 crc kubenswrapper[4805]: I1128 15:29:41.060638 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:29:41 crc kubenswrapper[4805]: I1128 15:29:41.060689 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:29:41 crc kubenswrapper[4805]: I1128 15:29:41.061286 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937"} pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:29:41 crc kubenswrapper[4805]: I1128 15:29:41.061591 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" containerID="cri-o://89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937" gracePeriod=600 Nov 28 15:29:41 crc kubenswrapper[4805]: E1128 15:29:41.119991 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 15:29:41 crc kubenswrapper[4805]: E1128 15:29:41.120206 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mlsq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7cqfs_openshift-marketplace(d53216ca-283e-4346-ace1-9fab11518bfd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:29:41 crc kubenswrapper[4805]: E1128 15:29:41.121639 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7cqfs" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" Nov 28 15:29:41 crc kubenswrapper[4805]: I1128 15:29:41.651932 4805 generic.go:334] "Generic (PLEG): container finished" podID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerID="89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937" exitCode=0 Nov 28 15:29:41 crc kubenswrapper[4805]: I1128 15:29:41.651981 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerDied","Data":"89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937"} Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.599671 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7cqfs" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.599938 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-tgr5n" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.703764 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.704127 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kw2q2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rgs2f_openshift-marketplace(c6b04434-a30b-4e22-8c5a-9cdaca124466): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.705380 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rgs2f" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.774212 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.774421 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gssxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dhzqc_openshift-marketplace(bf218599-d9d7-4e0e-a2ab-0949ded9fa58): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.775156 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.775309 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wlkbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-s2fzf_openshift-marketplace(eff8293e-510a-4cf5-afe2-5fefc4da0d3f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.776267 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dhzqc" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.777367 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-s2fzf" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.781277 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.781399 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-phqdp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-8d4l4_openshift-marketplace(db028335-3e13-44c6-844f-318168230b30): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:29:44 crc kubenswrapper[4805]: E1128 15:29:44.782534 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-8d4l4" podUID="db028335-3e13-44c6-844f-318168230b30" Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.056039 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 15:29:45 crc kubenswrapper[4805]: W1128 15:29:45.068506 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7779b1cc_4309_4cc8_b634_903dec64f766.slice/crio-8c9ffb65706cc348c53d02c964ff015f31a1b03ca9e4c1290afa3e5cf0f4b185 WatchSource:0}: Error finding container 8c9ffb65706cc348c53d02c964ff015f31a1b03ca9e4c1290afa3e5cf0f4b185: Status 404 returned error can't find the container with id 8c9ffb65706cc348c53d02c964ff015f31a1b03ca9e4c1290afa3e5cf0f4b185 Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.104788 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fplc8"] Nov 28 15:29:45 crc kubenswrapper[4805]: W1128 15:29:45.112872 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18cf186f_76f6_47bc_8db9_c9b7be3aaf09.slice/crio-c32157bf139af9542c32976a601c4ba432b7ed973011dcf9b25594a4144e2563 WatchSource:0}: Error finding container c32157bf139af9542c32976a601c4ba432b7ed973011dcf9b25594a4144e2563: Status 404 returned error can't find the container with id c32157bf139af9542c32976a601c4ba432b7ed973011dcf9b25594a4144e2563 Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.140930 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 15:29:45 crc kubenswrapper[4805]: W1128 15:29:45.159262 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod72212593_ffb8_48a1_a673_5532833d9134.slice/crio-50074013ee715f2c8cd7413abb613627f1cc3be1b1d365d3de87bfaad9d00d91 WatchSource:0}: Error finding container 50074013ee715f2c8cd7413abb613627f1cc3be1b1d365d3de87bfaad9d00d91: Status 404 returned error can't find the container with id 50074013ee715f2c8cd7413abb613627f1cc3be1b1d365d3de87bfaad9d00d91 Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.678796 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"72212593-ffb8-48a1-a673-5532833d9134","Type":"ContainerStarted","Data":"468ecd23b5a85c8997880aab4187cef05063a766c0d112399b0ea059b80a59e6"} Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.679702 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"72212593-ffb8-48a1-a673-5532833d9134","Type":"ContainerStarted","Data":"50074013ee715f2c8cd7413abb613627f1cc3be1b1d365d3de87bfaad9d00d91"} Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.680081 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fplc8" event={"ID":"18cf186f-76f6-47bc-8db9-c9b7be3aaf09","Type":"ContainerStarted","Data":"fff62984c1819b33bd7900b47f468d141ca862026f69c991977b004b34e359b3"} Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.680102 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fplc8" event={"ID":"18cf186f-76f6-47bc-8db9-c9b7be3aaf09","Type":"ContainerStarted","Data":"c32157bf139af9542c32976a601c4ba432b7ed973011dcf9b25594a4144e2563"} Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.681931 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"426a027662f6b8d978eced3586b5c9067232b166c02a8847de1e3b495b0e0a66"} Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.684030 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7779b1cc-4309-4cc8-b634-903dec64f766","Type":"ContainerStarted","Data":"5a74830945074fa6fc6404dadf22423964f7800c53990d6481bc7674a58728a4"} Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.684573 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7779b1cc-4309-4cc8-b634-903dec64f766","Type":"ContainerStarted","Data":"8c9ffb65706cc348c53d02c964ff015f31a1b03ca9e4c1290afa3e5cf0f4b185"} Nov 28 15:29:45 crc kubenswrapper[4805]: E1128 15:29:45.685132 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rgs2f" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" Nov 28 15:29:45 crc kubenswrapper[4805]: E1128 15:29:45.685565 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-8d4l4" podUID="db028335-3e13-44c6-844f-318168230b30" Nov 28 15:29:45 crc kubenswrapper[4805]: E1128 15:29:45.685949 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-s2fzf" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" Nov 28 15:29:45 crc kubenswrapper[4805]: E1128 15:29:45.686002 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dhzqc" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.698593 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=22.698574484 podStartE2EDuration="22.698574484s" podCreationTimestamp="2025-11-28 15:29:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:29:45.696302198 +0000 UTC m=+212.746093509" watchObservedRunningTime="2025-11-28 15:29:45.698574484 +0000 UTC m=+212.748365815" Nov 28 15:29:45 crc kubenswrapper[4805]: I1128 15:29:45.777615 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=27.777583131 podStartE2EDuration="27.777583131s" podCreationTimestamp="2025-11-28 15:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:29:45.775059198 +0000 UTC m=+212.824850549" watchObservedRunningTime="2025-11-28 15:29:45.777583131 +0000 UTC m=+212.827374442" Nov 28 15:29:46 crc kubenswrapper[4805]: I1128 15:29:46.691727 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fplc8" event={"ID":"18cf186f-76f6-47bc-8db9-c9b7be3aaf09","Type":"ContainerStarted","Data":"9292efbd52e0b6d768c099bf3dab27c28b1987285ab8fd3895fd688c7e37f57c"} Nov 28 15:29:46 crc kubenswrapper[4805]: I1128 15:29:46.713667 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-fplc8" podStartSLOduration=194.713644575 podStartE2EDuration="3m14.713644575s" podCreationTimestamp="2025-11-28 15:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:29:46.707312303 +0000 UTC m=+213.757103634" watchObservedRunningTime="2025-11-28 15:29:46.713644575 +0000 UTC m=+213.763435886" Nov 28 15:29:47 crc kubenswrapper[4805]: I1128 15:29:47.697986 4805 generic.go:334] "Generic (PLEG): container finished" podID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerID="82935b7ba460ce053a7e939956fdb891a539aa81a626fc7085c132043faf3c6c" exitCode=0 Nov 28 15:29:47 crc kubenswrapper[4805]: I1128 15:29:47.698257 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6fmm" event={"ID":"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07","Type":"ContainerDied","Data":"82935b7ba460ce053a7e939956fdb891a539aa81a626fc7085c132043faf3c6c"} Nov 28 15:29:47 crc kubenswrapper[4805]: I1128 15:29:47.702635 4805 generic.go:334] "Generic (PLEG): container finished" podID="7779b1cc-4309-4cc8-b634-903dec64f766" containerID="5a74830945074fa6fc6404dadf22423964f7800c53990d6481bc7674a58728a4" exitCode=0 Nov 28 15:29:47 crc kubenswrapper[4805]: I1128 15:29:47.702759 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7779b1cc-4309-4cc8-b634-903dec64f766","Type":"ContainerDied","Data":"5a74830945074fa6fc6404dadf22423964f7800c53990d6481bc7674a58728a4"} Nov 28 15:29:48 crc kubenswrapper[4805]: I1128 15:29:48.711767 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6fmm" event={"ID":"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07","Type":"ContainerStarted","Data":"c567e5a631e2408cceca537997282c14a19fe198f99f5c050e2fbce23b0adef2"} Nov 28 15:29:48 crc kubenswrapper[4805]: I1128 15:29:48.749195 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s6fmm" podStartSLOduration=2.7414515699999997 podStartE2EDuration="1m11.749179354s" podCreationTimestamp="2025-11-28 15:28:37 +0000 UTC" firstStartedPulling="2025-11-28 15:28:39.098461014 +0000 UTC m=+146.148252315" lastFinishedPulling="2025-11-28 15:29:48.106188788 +0000 UTC m=+215.155980099" observedRunningTime="2025-11-28 15:29:48.747154356 +0000 UTC m=+215.796945707" watchObservedRunningTime="2025-11-28 15:29:48.749179354 +0000 UTC m=+215.798970665" Nov 28 15:29:48 crc kubenswrapper[4805]: I1128 15:29:48.977585 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:48 crc kubenswrapper[4805]: I1128 15:29:48.996311 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7779b1cc-4309-4cc8-b634-903dec64f766-kubelet-dir\") pod \"7779b1cc-4309-4cc8-b634-903dec64f766\" (UID: \"7779b1cc-4309-4cc8-b634-903dec64f766\") " Nov 28 15:29:48 crc kubenswrapper[4805]: I1128 15:29:48.996411 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7779b1cc-4309-4cc8-b634-903dec64f766-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7779b1cc-4309-4cc8-b634-903dec64f766" (UID: "7779b1cc-4309-4cc8-b634-903dec64f766"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:29:48 crc kubenswrapper[4805]: I1128 15:29:48.996429 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7779b1cc-4309-4cc8-b634-903dec64f766-kube-api-access\") pod \"7779b1cc-4309-4cc8-b634-903dec64f766\" (UID: \"7779b1cc-4309-4cc8-b634-903dec64f766\") " Nov 28 15:29:48 crc kubenswrapper[4805]: I1128 15:29:48.997004 4805 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7779b1cc-4309-4cc8-b634-903dec64f766-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:49 crc kubenswrapper[4805]: I1128 15:29:49.003979 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7779b1cc-4309-4cc8-b634-903dec64f766-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7779b1cc-4309-4cc8-b634-903dec64f766" (UID: "7779b1cc-4309-4cc8-b634-903dec64f766"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:29:49 crc kubenswrapper[4805]: I1128 15:29:49.100821 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7779b1cc-4309-4cc8-b634-903dec64f766-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:49 crc kubenswrapper[4805]: I1128 15:29:49.717739 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7779b1cc-4309-4cc8-b634-903dec64f766","Type":"ContainerDied","Data":"8c9ffb65706cc348c53d02c964ff015f31a1b03ca9e4c1290afa3e5cf0f4b185"} Nov 28 15:29:49 crc kubenswrapper[4805]: I1128 15:29:49.718268 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c9ffb65706cc348c53d02c964ff015f31a1b03ca9e4c1290afa3e5cf0f4b185" Nov 28 15:29:49 crc kubenswrapper[4805]: I1128 15:29:49.717803 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 15:29:54 crc kubenswrapper[4805]: I1128 15:29:54.749685 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm6n6" event={"ID":"e9089841-40aa-479f-b56e-421629f24685","Type":"ContainerStarted","Data":"89da43cdca0e56e7dce566f21163b4306eb01a4d2248fc0731bf21b0696af09c"} Nov 28 15:29:55 crc kubenswrapper[4805]: I1128 15:29:55.757104 4805 generic.go:334] "Generic (PLEG): container finished" podID="e9089841-40aa-479f-b56e-421629f24685" containerID="89da43cdca0e56e7dce566f21163b4306eb01a4d2248fc0731bf21b0696af09c" exitCode=0 Nov 28 15:29:55 crc kubenswrapper[4805]: I1128 15:29:55.757156 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm6n6" event={"ID":"e9089841-40aa-479f-b56e-421629f24685","Type":"ContainerDied","Data":"89da43cdca0e56e7dce566f21163b4306eb01a4d2248fc0731bf21b0696af09c"} Nov 28 15:29:57 crc kubenswrapper[4805]: I1128 15:29:57.614277 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:29:57 crc kubenswrapper[4805]: I1128 15:29:57.616027 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:29:57 crc kubenswrapper[4805]: I1128 15:29:57.769589 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm6n6" event={"ID":"e9089841-40aa-479f-b56e-421629f24685","Type":"ContainerStarted","Data":"3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002"} Nov 28 15:29:57 crc kubenswrapper[4805]: I1128 15:29:57.788069 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hm6n6" podStartSLOduration=4.094665226 podStartE2EDuration="1m20.78805471s" podCreationTimestamp="2025-11-28 15:28:37 +0000 UTC" firstStartedPulling="2025-11-28 15:28:40.182613025 +0000 UTC m=+147.232404336" lastFinishedPulling="2025-11-28 15:29:56.876002509 +0000 UTC m=+223.925793820" observedRunningTime="2025-11-28 15:29:57.785747775 +0000 UTC m=+224.835539086" watchObservedRunningTime="2025-11-28 15:29:57.78805471 +0000 UTC m=+224.837846021" Nov 28 15:29:58 crc kubenswrapper[4805]: I1128 15:29:58.030474 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:29:58 crc kubenswrapper[4805]: I1128 15:29:58.030836 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:29:58 crc kubenswrapper[4805]: I1128 15:29:58.046914 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:29:58 crc kubenswrapper[4805]: I1128 15:29:58.815628 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:29:59 crc kubenswrapper[4805]: I1128 15:29:59.069091 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hm6n6" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="registry-server" probeResult="failure" output=< Nov 28 15:29:59 crc kubenswrapper[4805]: timeout: failed to connect service ":50051" within 1s Nov 28 15:29:59 crc kubenswrapper[4805]: > Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.133702 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr"] Nov 28 15:30:00 crc kubenswrapper[4805]: E1128 15:30:00.133969 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7779b1cc-4309-4cc8-b634-903dec64f766" containerName="pruner" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.133985 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7779b1cc-4309-4cc8-b634-903dec64f766" containerName="pruner" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.134121 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7779b1cc-4309-4cc8-b634-903dec64f766" containerName="pruner" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.134588 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.136975 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.137253 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.144825 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr"] Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.225459 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-secret-volume\") pod \"collect-profiles-29405730-b55nr\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.225543 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcsfz\" (UniqueName: \"kubernetes.io/projected/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-kube-api-access-pcsfz\") pod \"collect-profiles-29405730-b55nr\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.225618 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-config-volume\") pod \"collect-profiles-29405730-b55nr\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.327279 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-config-volume\") pod \"collect-profiles-29405730-b55nr\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.327715 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-secret-volume\") pod \"collect-profiles-29405730-b55nr\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.327772 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcsfz\" (UniqueName: \"kubernetes.io/projected/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-kube-api-access-pcsfz\") pod \"collect-profiles-29405730-b55nr\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.328219 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-config-volume\") pod \"collect-profiles-29405730-b55nr\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.336074 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-secret-volume\") pod \"collect-profiles-29405730-b55nr\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.342957 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcsfz\" (UniqueName: \"kubernetes.io/projected/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-kube-api-access-pcsfz\") pod \"collect-profiles-29405730-b55nr\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:00 crc kubenswrapper[4805]: I1128 15:30:00.450341 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.047587 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr"] Nov 28 15:30:07 crc kubenswrapper[4805]: W1128 15:30:07.054079 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd4e6f94_ea2a_4473_b799_fef63cb0e92d.slice/crio-c948b231ed7395b8e3d91987aa09e29cdb21efd9ec38b1bd9851c283d1257bdd WatchSource:0}: Error finding container c948b231ed7395b8e3d91987aa09e29cdb21efd9ec38b1bd9851c283d1257bdd: Status 404 returned error can't find the container with id c948b231ed7395b8e3d91987aa09e29cdb21efd9ec38b1bd9851c283d1257bdd Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.833540 4805 generic.go:334] "Generic (PLEG): container finished" podID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerID="f0fe943fd27b3d6f97690a00b062e43db356f7d7b8cb739b4c66f86f88d891dc" exitCode=0 Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.833592 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2fzf" event={"ID":"eff8293e-510a-4cf5-afe2-5fefc4da0d3f","Type":"ContainerDied","Data":"f0fe943fd27b3d6f97690a00b062e43db356f7d7b8cb739b4c66f86f88d891dc"} Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.838588 4805 generic.go:334] "Generic (PLEG): container finished" podID="db028335-3e13-44c6-844f-318168230b30" containerID="d3bb12c3972d097f73396516c8a9264bccff569673db36f991f940f635c24120" exitCode=0 Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.838686 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d4l4" event={"ID":"db028335-3e13-44c6-844f-318168230b30","Type":"ContainerDied","Data":"d3bb12c3972d097f73396516c8a9264bccff569673db36f991f940f635c24120"} Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.840719 4805 generic.go:334] "Generic (PLEG): container finished" podID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerID="92f2c1d521753f9522c867833b8127d99065dca037d8e999d81642db0e67c614" exitCode=0 Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.840811 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgr5n" event={"ID":"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa","Type":"ContainerDied","Data":"92f2c1d521753f9522c867833b8127d99065dca037d8e999d81642db0e67c614"} Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.851268 4805 generic.go:334] "Generic (PLEG): container finished" podID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerID="2609555d28210758274bef747b691dfc84dee650920cf8598677b2372d723115" exitCode=0 Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.851341 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgs2f" event={"ID":"c6b04434-a30b-4e22-8c5a-9cdaca124466","Type":"ContainerDied","Data":"2609555d28210758274bef747b691dfc84dee650920cf8598677b2372d723115"} Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.857807 4805 generic.go:334] "Generic (PLEG): container finished" podID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerID="972b766ed63a8c52bce906d61dfb675dbcb740b2de63e5b42c513d4ab09b8f25" exitCode=0 Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.857871 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhzqc" event={"ID":"bf218599-d9d7-4e0e-a2ab-0949ded9fa58","Type":"ContainerDied","Data":"972b766ed63a8c52bce906d61dfb675dbcb740b2de63e5b42c513d4ab09b8f25"} Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.860164 4805 generic.go:334] "Generic (PLEG): container finished" podID="d53216ca-283e-4346-ace1-9fab11518bfd" containerID="80f694736f365c5cd8b122ab1f63695286e9a5017b67971b706456fc727fb375" exitCode=0 Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.860209 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cqfs" event={"ID":"d53216ca-283e-4346-ace1-9fab11518bfd","Type":"ContainerDied","Data":"80f694736f365c5cd8b122ab1f63695286e9a5017b67971b706456fc727fb375"} Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.862734 4805 generic.go:334] "Generic (PLEG): container finished" podID="bd4e6f94-ea2a-4473-b799-fef63cb0e92d" containerID="863a80247214c8184f335bcfce3431900e0c53f1ce20bdb48ce0a14d6190cbe0" exitCode=0 Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.862773 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" event={"ID":"bd4e6f94-ea2a-4473-b799-fef63cb0e92d","Type":"ContainerDied","Data":"863a80247214c8184f335bcfce3431900e0c53f1ce20bdb48ce0a14d6190cbe0"} Nov 28 15:30:07 crc kubenswrapper[4805]: I1128 15:30:07.862820 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" event={"ID":"bd4e6f94-ea2a-4473-b799-fef63cb0e92d","Type":"ContainerStarted","Data":"c948b231ed7395b8e3d91987aa09e29cdb21efd9ec38b1bd9851c283d1257bdd"} Nov 28 15:30:08 crc kubenswrapper[4805]: I1128 15:30:08.080863 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:30:08 crc kubenswrapper[4805]: I1128 15:30:08.120397 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.252630 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.449665 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcsfz\" (UniqueName: \"kubernetes.io/projected/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-kube-api-access-pcsfz\") pod \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.449767 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-config-volume\") pod \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.449858 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-secret-volume\") pod \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\" (UID: \"bd4e6f94-ea2a-4473-b799-fef63cb0e92d\") " Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.450864 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-config-volume" (OuterVolumeSpecName: "config-volume") pod "bd4e6f94-ea2a-4473-b799-fef63cb0e92d" (UID: "bd4e6f94-ea2a-4473-b799-fef63cb0e92d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.459949 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bd4e6f94-ea2a-4473-b799-fef63cb0e92d" (UID: "bd4e6f94-ea2a-4473-b799-fef63cb0e92d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.459983 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-kube-api-access-pcsfz" (OuterVolumeSpecName: "kube-api-access-pcsfz") pod "bd4e6f94-ea2a-4473-b799-fef63cb0e92d" (UID: "bd4e6f94-ea2a-4473-b799-fef63cb0e92d"). InnerVolumeSpecName "kube-api-access-pcsfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.550821 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.550864 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.550877 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcsfz\" (UniqueName: \"kubernetes.io/projected/bd4e6f94-ea2a-4473-b799-fef63cb0e92d-kube-api-access-pcsfz\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.881752 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cqfs" event={"ID":"d53216ca-283e-4346-ace1-9fab11518bfd","Type":"ContainerStarted","Data":"5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0"} Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.883628 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" event={"ID":"bd4e6f94-ea2a-4473-b799-fef63cb0e92d","Type":"ContainerDied","Data":"c948b231ed7395b8e3d91987aa09e29cdb21efd9ec38b1bd9851c283d1257bdd"} Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.883646 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c948b231ed7395b8e3d91987aa09e29cdb21efd9ec38b1bd9851c283d1257bdd" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.883679 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-b55nr" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.891535 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2fzf" event={"ID":"eff8293e-510a-4cf5-afe2-5fefc4da0d3f","Type":"ContainerStarted","Data":"644cd402bc258a78446db53307622ef06b0c91d2e4bc740774eef3123ecb4ddc"} Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.893968 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgr5n" event={"ID":"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa","Type":"ContainerStarted","Data":"cd8b1626b1ca55f18bf08cf9481287374c0cadac06c385f2604e7eb505aafc75"} Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.895994 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgs2f" event={"ID":"c6b04434-a30b-4e22-8c5a-9cdaca124466","Type":"ContainerStarted","Data":"708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d"} Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.906350 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7cqfs" podStartSLOduration=4.537896927 podStartE2EDuration="1m35.906334427s" podCreationTimestamp="2025-11-28 15:28:34 +0000 UTC" firstStartedPulling="2025-11-28 15:28:36.97177364 +0000 UTC m=+144.021564951" lastFinishedPulling="2025-11-28 15:30:08.34021114 +0000 UTC m=+235.390002451" observedRunningTime="2025-11-28 15:30:09.90331264 +0000 UTC m=+236.953103961" watchObservedRunningTime="2025-11-28 15:30:09.906334427 +0000 UTC m=+236.956125738" Nov 28 15:30:09 crc kubenswrapper[4805]: I1128 15:30:09.950762 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rgs2f" podStartSLOduration=4.58670473 podStartE2EDuration="1m35.950740788s" podCreationTimestamp="2025-11-28 15:28:34 +0000 UTC" firstStartedPulling="2025-11-28 15:28:36.951051288 +0000 UTC m=+144.000842599" lastFinishedPulling="2025-11-28 15:30:08.315087346 +0000 UTC m=+235.364878657" observedRunningTime="2025-11-28 15:30:09.950159231 +0000 UTC m=+236.999950542" watchObservedRunningTime="2025-11-28 15:30:09.950740788 +0000 UTC m=+237.000532099" Nov 28 15:30:11 crc kubenswrapper[4805]: I1128 15:30:11.925949 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tgr5n" podStartSLOduration=4.884045992 podStartE2EDuration="1m35.925934197s" podCreationTimestamp="2025-11-28 15:28:36 +0000 UTC" firstStartedPulling="2025-11-28 15:28:38.083487469 +0000 UTC m=+145.133278780" lastFinishedPulling="2025-11-28 15:30:09.125375664 +0000 UTC m=+236.175166985" observedRunningTime="2025-11-28 15:30:11.923915438 +0000 UTC m=+238.973706749" watchObservedRunningTime="2025-11-28 15:30:11.925934197 +0000 UTC m=+238.975725508" Nov 28 15:30:12 crc kubenswrapper[4805]: I1128 15:30:12.186866 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hm6n6"] Nov 28 15:30:12 crc kubenswrapper[4805]: I1128 15:30:12.187127 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hm6n6" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="registry-server" containerID="cri-o://3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002" gracePeriod=2 Nov 28 15:30:12 crc kubenswrapper[4805]: I1128 15:30:12.930792 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s2fzf" podStartSLOduration=6.139571614 podStartE2EDuration="1m36.930774464s" podCreationTimestamp="2025-11-28 15:28:36 +0000 UTC" firstStartedPulling="2025-11-28 15:28:38.083496349 +0000 UTC m=+145.133287660" lastFinishedPulling="2025-11-28 15:30:08.874699159 +0000 UTC m=+235.924490510" observedRunningTime="2025-11-28 15:30:12.928635022 +0000 UTC m=+239.978426333" watchObservedRunningTime="2025-11-28 15:30:12.930774464 +0000 UTC m=+239.980565775" Nov 28 15:30:13 crc kubenswrapper[4805]: I1128 15:30:13.919607 4805 generic.go:334] "Generic (PLEG): container finished" podID="e9089841-40aa-479f-b56e-421629f24685" containerID="3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002" exitCode=0 Nov 28 15:30:13 crc kubenswrapper[4805]: I1128 15:30:13.919665 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm6n6" event={"ID":"e9089841-40aa-479f-b56e-421629f24685","Type":"ContainerDied","Data":"3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002"} Nov 28 15:30:14 crc kubenswrapper[4805]: I1128 15:30:14.800514 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:30:14 crc kubenswrapper[4805]: I1128 15:30:14.800569 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:30:14 crc kubenswrapper[4805]: I1128 15:30:14.858779 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:30:14 crc kubenswrapper[4805]: I1128 15:30:14.961515 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:30:14 crc kubenswrapper[4805]: I1128 15:30:14.976613 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:30:14 crc kubenswrapper[4805]: I1128 15:30:14.976701 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:30:15 crc kubenswrapper[4805]: I1128 15:30:15.019463 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:30:15 crc kubenswrapper[4805]: I1128 15:30:15.967940 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:30:16 crc kubenswrapper[4805]: I1128 15:30:16.578069 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:30:16 crc kubenswrapper[4805]: I1128 15:30:16.578118 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:30:16 crc kubenswrapper[4805]: I1128 15:30:16.637380 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:30:16 crc kubenswrapper[4805]: I1128 15:30:16.988538 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:30:17 crc kubenswrapper[4805]: I1128 15:30:17.037618 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:30:17 crc kubenswrapper[4805]: I1128 15:30:17.038204 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:30:17 crc kubenswrapper[4805]: I1128 15:30:17.082094 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:30:18 crc kubenswrapper[4805]: I1128 15:30:18.000926 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:30:18 crc kubenswrapper[4805]: E1128 15:30:18.031464 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002 is running failed: container process not found" containerID="3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 15:30:18 crc kubenswrapper[4805]: E1128 15:30:18.038501 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002 is running failed: container process not found" containerID="3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 15:30:18 crc kubenswrapper[4805]: E1128 15:30:18.039168 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002 is running failed: container process not found" containerID="3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 15:30:18 crc kubenswrapper[4805]: E1128 15:30:18.039259 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-hm6n6" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="registry-server" Nov 28 15:30:19 crc kubenswrapper[4805]: I1128 15:30:19.000040 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rgs2f"] Nov 28 15:30:19 crc kubenswrapper[4805]: I1128 15:30:19.000306 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rgs2f" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerName="registry-server" containerID="cri-o://708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d" gracePeriod=2 Nov 28 15:30:19 crc kubenswrapper[4805]: I1128 15:30:19.983898 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7cqfs"] Nov 28 15:30:19 crc kubenswrapper[4805]: I1128 15:30:19.984419 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7cqfs" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" containerName="registry-server" containerID="cri-o://5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0" gracePeriod=2 Nov 28 15:30:21 crc kubenswrapper[4805]: I1128 15:30:21.385862 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgr5n"] Nov 28 15:30:21 crc kubenswrapper[4805]: I1128 15:30:21.386435 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tgr5n" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerName="registry-server" containerID="cri-o://cd8b1626b1ca55f18bf08cf9481287374c0cadac06c385f2604e7eb505aafc75" gracePeriod=2 Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.524333 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.679623 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxpjb\" (UniqueName: \"kubernetes.io/projected/e9089841-40aa-479f-b56e-421629f24685-kube-api-access-sxpjb\") pod \"e9089841-40aa-479f-b56e-421629f24685\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.680240 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-catalog-content\") pod \"e9089841-40aa-479f-b56e-421629f24685\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.680281 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-utilities\") pod \"e9089841-40aa-479f-b56e-421629f24685\" (UID: \"e9089841-40aa-479f-b56e-421629f24685\") " Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.692689 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-utilities" (OuterVolumeSpecName: "utilities") pod "e9089841-40aa-479f-b56e-421629f24685" (UID: "e9089841-40aa-479f-b56e-421629f24685"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.695200 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9089841-40aa-479f-b56e-421629f24685-kube-api-access-sxpjb" (OuterVolumeSpecName: "kube-api-access-sxpjb") pod "e9089841-40aa-479f-b56e-421629f24685" (UID: "e9089841-40aa-479f-b56e-421629f24685"). InnerVolumeSpecName "kube-api-access-sxpjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.783117 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxpjb\" (UniqueName: \"kubernetes.io/projected/e9089841-40aa-479f-b56e-421629f24685-kube-api-access-sxpjb\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.783154 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.970616 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7cqfs_d53216ca-283e-4346-ace1-9fab11518bfd/registry-server/0.log" Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.971369 4805 generic.go:334] "Generic (PLEG): container finished" podID="d53216ca-283e-4346-ace1-9fab11518bfd" containerID="5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0" exitCode=137 Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.971439 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cqfs" event={"ID":"d53216ca-283e-4346-ace1-9fab11518bfd","Type":"ContainerDied","Data":"5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0"} Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.973310 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm6n6" event={"ID":"e9089841-40aa-479f-b56e-421629f24685","Type":"ContainerDied","Data":"ffdf9858dcc79eff07c7aa05400f1a16d6e641ec0b13ff4a08b3bdc1149fcf6d"} Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.973353 4805 scope.go:117] "RemoveContainer" containerID="3faed9d0caa241905ad99ba3e7c0dfa119b8c4505f6dcf1a1a624934d90b7002" Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.973511 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hm6n6" Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.976088 4805 generic.go:334] "Generic (PLEG): container finished" podID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerID="cd8b1626b1ca55f18bf08cf9481287374c0cadac06c385f2604e7eb505aafc75" exitCode=0 Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.976131 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgr5n" event={"ID":"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa","Type":"ContainerDied","Data":"cd8b1626b1ca55f18bf08cf9481287374c0cadac06c385f2604e7eb505aafc75"} Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.980037 4805 generic.go:334] "Generic (PLEG): container finished" podID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerID="708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d" exitCode=0 Nov 28 15:30:22 crc kubenswrapper[4805]: I1128 15:30:22.980079 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgs2f" event={"ID":"c6b04434-a30b-4e22-8c5a-9cdaca124466","Type":"ContainerDied","Data":"708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d"} Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.234068 4805 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.234350 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="extract-content" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.234381 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="extract-content" Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.234408 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="extract-utilities" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.234417 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="extract-utilities" Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.234428 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="registry-server" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.234436 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="registry-server" Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.234447 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4e6f94-ea2a-4473-b799-fef63cb0e92d" containerName="collect-profiles" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.234454 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4e6f94-ea2a-4473-b799-fef63cb0e92d" containerName="collect-profiles" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.234580 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd4e6f94-ea2a-4473-b799-fef63cb0e92d" containerName="collect-profiles" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.234594 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9089841-40aa-479f-b56e-421629f24685" containerName="registry-server" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.235028 4805 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.235249 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.235453 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1" gracePeriod=15 Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.235480 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78" gracePeriod=15 Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.235511 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826" gracePeriod=15 Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.235551 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781" gracePeriod=15 Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.235532 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26" gracePeriod=15 Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.235712 4805 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.235983 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.235997 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.236007 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236014 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.236025 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236033 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.236044 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236051 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.236062 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236072 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.236090 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236097 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 15:30:23 crc kubenswrapper[4805]: E1128 15:30:23.236110 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236118 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236249 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236261 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236270 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236283 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236298 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.236529 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.271937 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.288833 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.288907 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.288941 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.288982 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.289006 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.289036 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.289065 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.289096 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389476 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389504 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389532 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389548 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389616 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389624 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389666 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389659 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389680 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389648 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389833 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389911 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.389974 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.390030 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.391825 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.391928 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.506742 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9089841-40aa-479f-b56e-421629f24685" (UID: "e9089841-40aa-479f-b56e-421629f24685"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.566846 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.591848 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.594452 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9089841-40aa-479f-b56e-421629f24685-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.594404 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.595429 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.987711 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.990444 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.997286 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26" exitCode=2 Nov 28 15:30:23 crc kubenswrapper[4805]: I1128 15:30:23.999142 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhzqc" event={"ID":"bf218599-d9d7-4e0e-a2ab-0949ded9fa58","Type":"ContainerStarted","Data":"eb15017ccdf744b374fcdaec4b8aea1b2514300d65bf7b400c26f17a7a53f4fd"} Nov 28 15:30:24 crc kubenswrapper[4805]: E1128 15:30:24.801398 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d is running failed: container process not found" containerID="708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 15:30:24 crc kubenswrapper[4805]: E1128 15:30:24.801803 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d is running failed: container process not found" containerID="708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 15:30:24 crc kubenswrapper[4805]: E1128 15:30:24.802078 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d is running failed: container process not found" containerID="708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 15:30:24 crc kubenswrapper[4805]: E1128 15:30:24.802121 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-rgs2f" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerName="registry-server" Nov 28 15:30:24 crc kubenswrapper[4805]: E1128 15:30:24.977330 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0 is running failed: container process not found" containerID="5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 15:30:24 crc kubenswrapper[4805]: E1128 15:30:24.977822 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0 is running failed: container process not found" containerID="5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 15:30:24 crc kubenswrapper[4805]: E1128 15:30:24.978175 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0 is running failed: container process not found" containerID="5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 15:30:24 crc kubenswrapper[4805]: E1128 15:30:24.978211 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-7cqfs" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" containerName="registry-server" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.758307 4805 scope.go:117] "RemoveContainer" containerID="89da43cdca0e56e7dce566f21163b4306eb01a4d2248fc0731bf21b0696af09c" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.914703 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7cqfs_d53216ca-283e-4346-ace1-9fab11518bfd/registry-server/0.log" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.915724 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.916137 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.916442 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.916818 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.917571 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.918013 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.918288 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.918660 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.918897 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.923103 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.923740 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.924247 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.924692 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.925040 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.925469 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:25 crc kubenswrapper[4805]: I1128 15:30:25.943092 4805 scope.go:117] "RemoveContainer" containerID="cbaa3ce5d78508340baff32d7fffeab7879c9eae9ca3ef66f4a9da550fa9a8c2" Nov 28 15:30:25 crc kubenswrapper[4805]: W1128 15:30:25.946388 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-01e64e2e1269bb8f946728af300cd6444a96f8e1967a57c58c975d3c64909b14 WatchSource:0}: Error finding container 01e64e2e1269bb8f946728af300cd6444a96f8e1967a57c58c975d3c64909b14: Status 404 returned error can't find the container with id 01e64e2e1269bb8f946728af300cd6444a96f8e1967a57c58c975d3c64909b14 Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.009173 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"01e64e2e1269bb8f946728af300cd6444a96f8e1967a57c58c975d3c64909b14"} Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.012053 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgr5n" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.012046 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgr5n" event={"ID":"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa","Type":"ContainerDied","Data":"9f4b512043c98bcced425657cb2268c5affabd65f19d2539a3d3c0e24b8b841d"} Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.012760 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.013238 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.013590 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.014018 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.014416 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.014925 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgs2f" event={"ID":"c6b04434-a30b-4e22-8c5a-9cdaca124466","Type":"ContainerDied","Data":"00319e1ad1bf63777319714659b1b6f01e340948d65aec8acb8a3db4d3be68a2"} Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.014931 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgs2f" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.015782 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.016097 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.016414 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.016597 4805 generic.go:334] "Generic (PLEG): container finished" podID="72212593-ffb8-48a1-a673-5532833d9134" containerID="468ecd23b5a85c8997880aab4187cef05063a766c0d112399b0ea059b80a59e6" exitCode=0 Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.016665 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"72212593-ffb8-48a1-a673-5532833d9134","Type":"ContainerDied","Data":"468ecd23b5a85c8997880aab4187cef05063a766c0d112399b0ea059b80a59e6"} Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.016663 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.016949 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.017297 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.017778 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.018060 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.018297 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.018545 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.018787 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.020045 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-utilities\") pod \"d53216ca-283e-4346-ace1-9fab11518bfd\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.020079 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-utilities\") pod \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.020102 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l78sb\" (UniqueName: \"kubernetes.io/projected/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-kube-api-access-l78sb\") pod \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.020161 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-catalog-content\") pod \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\" (UID: \"7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.020211 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mlsq\" (UniqueName: \"kubernetes.io/projected/d53216ca-283e-4346-ace1-9fab11518bfd-kube-api-access-7mlsq\") pod \"d53216ca-283e-4346-ace1-9fab11518bfd\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.020248 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-catalog-content\") pod \"d53216ca-283e-4346-ace1-9fab11518bfd\" (UID: \"d53216ca-283e-4346-ace1-9fab11518bfd\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.020821 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-utilities" (OuterVolumeSpecName: "utilities") pod "7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" (UID: "7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.020910 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-utilities" (OuterVolumeSpecName: "utilities") pod "d53216ca-283e-4346-ace1-9fab11518bfd" (UID: "d53216ca-283e-4346-ace1-9fab11518bfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.023249 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d53216ca-283e-4346-ace1-9fab11518bfd-kube-api-access-7mlsq" (OuterVolumeSpecName: "kube-api-access-7mlsq") pod "d53216ca-283e-4346-ace1-9fab11518bfd" (UID: "d53216ca-283e-4346-ace1-9fab11518bfd"). InnerVolumeSpecName "kube-api-access-7mlsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.024058 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-kube-api-access-l78sb" (OuterVolumeSpecName: "kube-api-access-l78sb") pod "7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" (UID: "7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa"). InnerVolumeSpecName "kube-api-access-l78sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.025052 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7cqfs_d53216ca-283e-4346-ace1-9fab11518bfd/registry-server/0.log" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.025897 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cqfs" event={"ID":"d53216ca-283e-4346-ace1-9fab11518bfd","Type":"ContainerDied","Data":"9f59f8b0e43942fa7d88be3d07651c9215957a017d73a3bfa1a4bcd31a5373e7"} Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.025997 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7cqfs" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.026693 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.026848 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.027006 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.027223 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.028784 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.028955 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.033459 4805 scope.go:117] "RemoveContainer" containerID="cd8b1626b1ca55f18bf08cf9481287374c0cadac06c385f2604e7eb505aafc75" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.037383 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.041180 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.042132 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.042578 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.042784 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78" exitCode=0 Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.042808 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781" exitCode=0 Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.042815 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826" exitCode=0 Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.042824 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1" exitCode=0 Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.042878 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c24723337e2925ea47cd9132820dd4a5d9b791a389c306f36c433bc130607c27" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.043358 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.043814 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.044063 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.044475 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.046032 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.046340 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.047194 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.047432 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.047637 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.047889 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.048106 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.048349 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.048736 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.049107 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.049313 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.049559 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.050195 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" (UID: "7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.065217 4805 scope.go:117] "RemoveContainer" containerID="92f2c1d521753f9522c867833b8127d99065dca037d8e999d81642db0e67c614" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.075045 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d53216ca-283e-4346-ace1-9fab11518bfd" (UID: "d53216ca-283e-4346-ace1-9fab11518bfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.083037 4805 scope.go:117] "RemoveContainer" containerID="9d477a9f7cf06a01e0bda14061263382874f8c57d67d0beeeeb44f4b6a1d4e60" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.111242 4805 scope.go:117] "RemoveContainer" containerID="708d92ae2335c87a23eaf71aece1efedfe6ac00ba0c5445d1e7f144cbdf5689d" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122024 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-utilities\") pod \"c6b04434-a30b-4e22-8c5a-9cdaca124466\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122133 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-catalog-content\") pod \"c6b04434-a30b-4e22-8c5a-9cdaca124466\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122275 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122402 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122432 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw2q2\" (UniqueName: \"kubernetes.io/projected/c6b04434-a30b-4e22-8c5a-9cdaca124466-kube-api-access-kw2q2\") pod \"c6b04434-a30b-4e22-8c5a-9cdaca124466\" (UID: \"c6b04434-a30b-4e22-8c5a-9cdaca124466\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122479 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122669 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122779 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122801 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-utilities" (OuterVolumeSpecName: "utilities") pod "c6b04434-a30b-4e22-8c5a-9cdaca124466" (UID: "c6b04434-a30b-4e22-8c5a-9cdaca124466"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122839 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122942 4805 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122959 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122970 4805 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.122995 4805 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.123005 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mlsq\" (UniqueName: \"kubernetes.io/projected/d53216ca-283e-4346-ace1-9fab11518bfd-kube-api-access-7mlsq\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.123015 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.123024 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.123034 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53216ca-283e-4346-ace1-9fab11518bfd-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.123044 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.123054 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l78sb\" (UniqueName: \"kubernetes.io/projected/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa-kube-api-access-l78sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.126486 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b04434-a30b-4e22-8c5a-9cdaca124466-kube-api-access-kw2q2" (OuterVolumeSpecName: "kube-api-access-kw2q2") pod "c6b04434-a30b-4e22-8c5a-9cdaca124466" (UID: "c6b04434-a30b-4e22-8c5a-9cdaca124466"). InnerVolumeSpecName "kube-api-access-kw2q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.130707 4805 scope.go:117] "RemoveContainer" containerID="2609555d28210758274bef747b691dfc84dee650920cf8598677b2372d723115" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.148581 4805 scope.go:117] "RemoveContainer" containerID="b2d20787fca4d3655496a5b84092c393da251ee288de74552d41e4d96e404918" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.163326 4805 scope.go:117] "RemoveContainer" containerID="5b9464e8c822013ee86858d36625bbaa3fcbb4bd8a48c7e731758ce494570dc0" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.185629 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6b04434-a30b-4e22-8c5a-9cdaca124466" (UID: "c6b04434-a30b-4e22-8c5a-9cdaca124466"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.186539 4805 scope.go:117] "RemoveContainer" containerID="80f694736f365c5cd8b122ab1f63695286e9a5017b67971b706456fc727fb375" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.207095 4805 scope.go:117] "RemoveContainer" containerID="d7c90d015f58ac95ab50438f250cd78877ac98fe6b856cf8362686840dd945fe" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.224416 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b04434-a30b-4e22-8c5a-9cdaca124466-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.224738 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw2q2\" (UniqueName: \"kubernetes.io/projected/c6b04434-a30b-4e22-8c5a-9cdaca124466-kube-api-access-kw2q2\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.226264 4805 scope.go:117] "RemoveContainer" containerID="b93cdb445912a5ee3559632c8d3d5fc1020a6c968087188db91730924c0b8a68" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.326154 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.326360 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.327147 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.327323 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.327532 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.327721 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.327917 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.328153 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.333632 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.333902 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.334313 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.334474 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.334613 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.334756 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.334896 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.335034 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.341768 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.342095 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.342288 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.342869 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.343078 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.343310 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.343525 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:26 crc kubenswrapper[4805]: I1128 15:30:26.343706 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.054326 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.055230 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.056818 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.057168 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.057489 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.057737 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.057945 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.058112 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.058289 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.058521 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.060415 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d4l4" event={"ID":"db028335-3e13-44c6-844f-318168230b30","Type":"ContainerStarted","Data":"fb2dcff113a351d2085a3e177008f9bcc2129e6e43d6b47f4219a16fd74664ae"} Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.061022 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.061288 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.061465 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.061655 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.062882 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835"} Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.064503 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.064721 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.064927 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.065115 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.065333 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.065644 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.065826 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.065981 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.066157 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.066348 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.066529 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.066745 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.066968 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.067193 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.090808 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.091433 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.091839 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.092067 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.092335 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.093389 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.093631 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.096431 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.096880 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: E1128 15:30:27.203943 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:30:27Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:30:27Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:30:27Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T15:30:27Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:20434c856c20158a4c73986bf7de93188afa338ed356d293a59f9e621072cfc3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:24f7dab5f4a6fcbb16d41b8a7345f9f9bae2ef1e2c53abed71c4f18eeafebc85\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1605131077},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:1ab7704f67839bb3705d0c80bea6f7197f233d472860c3005433c90d7786dd54\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:9c13035c7ccf9d13a21c9219d8d0d462fa2fdb4fe128d9724443784b1ed9a318\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1205801806},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:485eae41e5a1129e031da03a9bc899702d16da22589d58a8e0c2910bc0226a23\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:86681c5c7f102911ba70f243ae7524f9a76939abbb50c93b1c80b70e07ccba62\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1195438934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:42da3a01b99987f17824a70b0ac9cde8d27a0ea39d325b9b7216ebdc5ba1f406\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:8932ced4defd2733d4740ea31dd7a6050447207c72233491a6ffdb06926137e7\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1122761533},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: E1128 15:30:27.204654 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: E1128 15:30:27.204866 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: E1128 15:30:27.205059 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: E1128 15:30:27.205445 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: E1128 15:30:27.205479 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.212197 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.362826 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.363979 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.364271 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.364533 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.364751 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.364958 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.365247 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.365459 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.365657 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.541347 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-var-lock\") pod \"72212593-ffb8-48a1-a673-5532833d9134\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.541419 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-kubelet-dir\") pod \"72212593-ffb8-48a1-a673-5532833d9134\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.541551 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "72212593-ffb8-48a1-a673-5532833d9134" (UID: "72212593-ffb8-48a1-a673-5532833d9134"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.541576 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-var-lock" (OuterVolumeSpecName: "var-lock") pod "72212593-ffb8-48a1-a673-5532833d9134" (UID: "72212593-ffb8-48a1-a673-5532833d9134"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.541612 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72212593-ffb8-48a1-a673-5532833d9134-kube-api-access\") pod \"72212593-ffb8-48a1-a673-5532833d9134\" (UID: \"72212593-ffb8-48a1-a673-5532833d9134\") " Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.542104 4805 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.542126 4805 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72212593-ffb8-48a1-a673-5532833d9134-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.548801 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72212593-ffb8-48a1-a673-5532833d9134-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "72212593-ffb8-48a1-a673-5532833d9134" (UID: "72212593-ffb8-48a1-a673-5532833d9134"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:30:27 crc kubenswrapper[4805]: I1128 15:30:27.644152 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72212593-ffb8-48a1-a673-5532833d9134-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.075233 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"72212593-ffb8-48a1-a673-5532833d9134","Type":"ContainerDied","Data":"50074013ee715f2c8cd7413abb613627f1cc3be1b1d365d3de87bfaad9d00d91"} Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.075270 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50074013ee715f2c8cd7413abb613627f1cc3be1b1d365d3de87bfaad9d00d91" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.075421 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.086785 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.087143 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.087659 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.087907 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.088183 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.088533 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.088824 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.089198 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: E1128 15:30:28.395056 4805 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.196:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-dhzqc.187c355c03a1737a openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-dhzqc,UID:bf218599-d9d7-4e0e-a2ab-0949ded9fa58,APIVersion:v1,ResourceVersion:28350,FieldPath:spec.containers{registry-server},},Reason:Created,Message:Created container registry-server,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 15:30:23.336567674 +0000 UTC m=+250.386358975,LastTimestamp:2025-11-28 15:30:23.336567674 +0000 UTC m=+250.386358975,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 15:30:28 crc kubenswrapper[4805]: E1128 15:30:28.427211 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: E1128 15:30:28.427576 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: E1128 15:30:28.427968 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: E1128 15:30:28.428549 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: E1128 15:30:28.428946 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:28 crc kubenswrapper[4805]: I1128 15:30:28.428977 4805 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 28 15:30:28 crc kubenswrapper[4805]: E1128 15:30:28.429208 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="200ms" Nov 28 15:30:28 crc kubenswrapper[4805]: E1128 15:30:28.629867 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="400ms" Nov 28 15:30:29 crc kubenswrapper[4805]: E1128 15:30:29.031167 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="800ms" Nov 28 15:30:29 crc kubenswrapper[4805]: E1128 15:30:29.832322 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="1.6s" Nov 28 15:30:31 crc kubenswrapper[4805]: E1128 15:30:31.433697 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="3.2s" Nov 28 15:30:33 crc kubenswrapper[4805]: I1128 15:30:33.207307 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:33 crc kubenswrapper[4805]: I1128 15:30:33.207964 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:33 crc kubenswrapper[4805]: I1128 15:30:33.208547 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:33 crc kubenswrapper[4805]: I1128 15:30:33.209053 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:33 crc kubenswrapper[4805]: I1128 15:30:33.209871 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:33 crc kubenswrapper[4805]: I1128 15:30:33.210242 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:33 crc kubenswrapper[4805]: I1128 15:30:33.210726 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:33 crc kubenswrapper[4805]: I1128 15:30:33.211001 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.203293 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.204217 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.204823 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.205295 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.205577 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.205812 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.206118 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.206381 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.206589 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.217538 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.217573 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:34 crc kubenswrapper[4805]: E1128 15:30:34.217944 4805 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.218351 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:34 crc kubenswrapper[4805]: W1128 15:30:34.244110 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-47d9c506f1dd423bb8b9e678afeb89a5b4ea71ef4705d984257d4d90e8e16911 WatchSource:0}: Error finding container 47d9c506f1dd423bb8b9e678afeb89a5b4ea71ef4705d984257d4d90e8e16911: Status 404 returned error can't find the container with id 47d9c506f1dd423bb8b9e678afeb89a5b4ea71ef4705d984257d4d90e8e16911 Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.403809 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.404854 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.456794 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.457785 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.458124 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.458353 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.458570 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.458767 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.458964 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.459154 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: I1128 15:30:34.459346 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:34 crc kubenswrapper[4805]: E1128 15:30:34.635022 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.196:6443: connect: connection refused" interval="6.4s" Nov 28 15:30:35 crc kubenswrapper[4805]: E1128 15:30:35.064068 4805 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.196:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-dhzqc.187c355c03a1737a openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-dhzqc,UID:bf218599-d9d7-4e0e-a2ab-0949ded9fa58,APIVersion:v1,ResourceVersion:28350,FieldPath:spec.containers{registry-server},},Reason:Created,Message:Created container registry-server,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 15:30:23.336567674 +0000 UTC m=+250.386358975,LastTimestamp:2025-11-28 15:30:23.336567674 +0000 UTC m=+250.386358975,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.114344 4805 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="f2feedafe8dcfccfa7c5d1c1dcab4614276de26fcf0bd7b6046e811f45fe7660" exitCode=0 Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.114823 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"f2feedafe8dcfccfa7c5d1c1dcab4614276de26fcf0bd7b6046e811f45fe7660"} Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.114883 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"47d9c506f1dd423bb8b9e678afeb89a5b4ea71ef4705d984257d4d90e8e16911"} Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.115211 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.115237 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.115641 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: E1128 15:30:35.115650 4805 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.116069 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.116351 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.116775 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.117096 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.117374 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.117873 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.118350 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.157098 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.158022 4805 status_manager.go:851] "Failed to get status for pod" podUID="72212593-ffb8-48a1-a673-5532833d9134" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.158510 4805 status_manager.go:851] "Failed to get status for pod" podUID="e9089841-40aa-479f-b56e-421629f24685" pod="openshift-marketplace/redhat-operators-hm6n6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hm6n6\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.158831 4805 status_manager.go:851] "Failed to get status for pod" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" pod="openshift-marketplace/community-operators-rgs2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rgs2f\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.159150 4805 status_manager.go:851] "Failed to get status for pod" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" pod="openshift-marketplace/certified-operators-7cqfs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-7cqfs\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.159543 4805 status_manager.go:851] "Failed to get status for pod" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" pod="openshift-marketplace/redhat-marketplace-tgr5n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-tgr5n\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.159834 4805 status_manager.go:851] "Failed to get status for pod" podUID="db028335-3e13-44c6-844f-318168230b30" pod="openshift-marketplace/community-operators-8d4l4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-8d4l4\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.160087 4805 status_manager.go:851] "Failed to get status for pod" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" pod="openshift-marketplace/certified-operators-dhzqc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-dhzqc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.160392 4805 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.196:6443: connect: connection refused" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.522625 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.522851 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:30:35 crc kubenswrapper[4805]: I1128 15:30:35.573246 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:30:36 crc kubenswrapper[4805]: I1128 15:30:36.124571 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d34605e9fec7cec9d05c6fab35f9f383f744649852c312300d2e1479c9875db4"} Nov 28 15:30:36 crc kubenswrapper[4805]: I1128 15:30:36.124928 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ec815045effc40e089ada7a95543db8f703b876312166c381dfe20d69a57908c"} Nov 28 15:30:36 crc kubenswrapper[4805]: I1128 15:30:36.124953 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"de8bb897496b9328620ed994955678f6b15174f991cdc5494aea0d7d6bc298de"} Nov 28 15:30:36 crc kubenswrapper[4805]: I1128 15:30:36.188095 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:30:37 crc kubenswrapper[4805]: I1128 15:30:37.132089 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9c12bd41fae8acec7a2a40eb84aaca04ce40f3683d55218d25ca990fe65c6a94"} Nov 28 15:30:37 crc kubenswrapper[4805]: I1128 15:30:37.132402 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2f649c519b05031768cf27ab8f0314ceb410cd412388d1a059fc1ef1606f30c5"} Nov 28 15:30:37 crc kubenswrapper[4805]: I1128 15:30:37.132423 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:37 crc kubenswrapper[4805]: I1128 15:30:37.132446 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:37 crc kubenswrapper[4805]: I1128 15:30:37.132470 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:38 crc kubenswrapper[4805]: I1128 15:30:38.140240 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 15:30:38 crc kubenswrapper[4805]: I1128 15:30:38.140287 4805 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18" exitCode=1 Nov 28 15:30:38 crc kubenswrapper[4805]: I1128 15:30:38.140409 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18"} Nov 28 15:30:38 crc kubenswrapper[4805]: I1128 15:30:38.141050 4805 scope.go:117] "RemoveContainer" containerID="a26664b63487e2131c0e331e731d88b1f50e8af629fad69f47bd975086cb2b18" Nov 28 15:30:38 crc kubenswrapper[4805]: I1128 15:30:38.929469 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:30:39 crc kubenswrapper[4805]: I1128 15:30:39.150499 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 15:30:39 crc kubenswrapper[4805]: I1128 15:30:39.151611 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"db1984582b413da58f946bdfcf96c2144a8a606dc47c8f4ae50f60c9e44e2fd7"} Nov 28 15:30:39 crc kubenswrapper[4805]: I1128 15:30:39.219501 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:39 crc kubenswrapper[4805]: I1128 15:30:39.219548 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:39 crc kubenswrapper[4805]: I1128 15:30:39.227044 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:40 crc kubenswrapper[4805]: I1128 15:30:40.377436 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:30:42 crc kubenswrapper[4805]: I1128 15:30:42.143502 4805 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:42 crc kubenswrapper[4805]: I1128 15:30:42.173813 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:42 crc kubenswrapper[4805]: I1128 15:30:42.173851 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:42 crc kubenswrapper[4805]: I1128 15:30:42.178617 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:43 crc kubenswrapper[4805]: I1128 15:30:43.177685 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:43 crc kubenswrapper[4805]: I1128 15:30:43.178000 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8f42d3-5dd5-49c9-8645-382075c8e289" Nov 28 15:30:43 crc kubenswrapper[4805]: I1128 15:30:43.220138 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="dbc8a4ac-496e-4a3f-b2fb-777b4b2dc345" Nov 28 15:30:47 crc kubenswrapper[4805]: I1128 15:30:47.929671 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:30:47 crc kubenswrapper[4805]: I1128 15:30:47.935636 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:30:48 crc kubenswrapper[4805]: I1128 15:30:48.216961 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 15:30:49 crc kubenswrapper[4805]: I1128 15:30:49.608435 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 15:30:49 crc kubenswrapper[4805]: I1128 15:30:49.920619 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 15:30:50 crc kubenswrapper[4805]: I1128 15:30:50.097026 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 15:30:50 crc kubenswrapper[4805]: I1128 15:30:50.179282 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 15:30:50 crc kubenswrapper[4805]: I1128 15:30:50.208706 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 15:30:50 crc kubenswrapper[4805]: I1128 15:30:50.578125 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 15:30:50 crc kubenswrapper[4805]: I1128 15:30:50.981603 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 15:30:51 crc kubenswrapper[4805]: I1128 15:30:51.135130 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 15:30:51 crc kubenswrapper[4805]: I1128 15:30:51.258907 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 15:30:51 crc kubenswrapper[4805]: I1128 15:30:51.371476 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 15:30:51 crc kubenswrapper[4805]: I1128 15:30:51.408119 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 15:30:51 crc kubenswrapper[4805]: I1128 15:30:51.588198 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 15:30:51 crc kubenswrapper[4805]: I1128 15:30:51.734614 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 15:30:51 crc kubenswrapper[4805]: I1128 15:30:51.751320 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 15:30:51 crc kubenswrapper[4805]: I1128 15:30:51.779908 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 15:30:52 crc kubenswrapper[4805]: I1128 15:30:52.095558 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 15:30:52 crc kubenswrapper[4805]: I1128 15:30:52.339432 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 15:30:52 crc kubenswrapper[4805]: I1128 15:30:52.816618 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 15:30:52 crc kubenswrapper[4805]: I1128 15:30:52.820480 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 15:30:52 crc kubenswrapper[4805]: I1128 15:30:52.844134 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 15:30:52 crc kubenswrapper[4805]: I1128 15:30:52.874510 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 15:30:53 crc kubenswrapper[4805]: I1128 15:30:53.484182 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 15:30:53 crc kubenswrapper[4805]: I1128 15:30:53.784451 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 15:30:53 crc kubenswrapper[4805]: I1128 15:30:53.832782 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 15:30:53 crc kubenswrapper[4805]: I1128 15:30:53.900478 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 15:30:54 crc kubenswrapper[4805]: I1128 15:30:54.251160 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 15:30:54 crc kubenswrapper[4805]: I1128 15:30:54.548787 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 15:30:54 crc kubenswrapper[4805]: I1128 15:30:54.612050 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 15:30:54 crc kubenswrapper[4805]: I1128 15:30:54.962543 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 15:30:55 crc kubenswrapper[4805]: I1128 15:30:55.135749 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 15:30:55 crc kubenswrapper[4805]: I1128 15:30:55.161641 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 15:30:55 crc kubenswrapper[4805]: I1128 15:30:55.362935 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 15:30:55 crc kubenswrapper[4805]: I1128 15:30:55.385399 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 15:30:55 crc kubenswrapper[4805]: I1128 15:30:55.438697 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 15:30:55 crc kubenswrapper[4805]: I1128 15:30:55.807844 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 15:30:55 crc kubenswrapper[4805]: I1128 15:30:55.836820 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 15:30:55 crc kubenswrapper[4805]: I1128 15:30:55.894648 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 15:30:55 crc kubenswrapper[4805]: I1128 15:30:55.976316 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.035770 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.056585 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.095942 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.469128 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.480589 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.525834 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.559311 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.650512 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.719902 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 15:30:56 crc kubenswrapper[4805]: I1128 15:30:56.751015 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.075813 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.173770 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.244929 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.276314 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.364895 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.406705 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.418954 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.478451 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.494012 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.544775 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.564844 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.611236 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.636460 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.639752 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.713105 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.787970 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.797128 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.829331 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.830392 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.873622 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.883861 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.900017 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 15:30:57 crc kubenswrapper[4805]: I1128 15:30:57.987161 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.007356 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.050301 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.059312 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.204634 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.312441 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.394683 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.416900 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.452157 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.645566 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.654422 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.671463 4805 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.673311 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dhzqc" podStartSLOduration=39.432372954 podStartE2EDuration="2m24.673281988s" podCreationTimestamp="2025-11-28 15:28:34 +0000 UTC" firstStartedPulling="2025-11-28 15:28:36.965757489 +0000 UTC m=+144.015548800" lastFinishedPulling="2025-11-28 15:30:22.206666503 +0000 UTC m=+249.256457834" observedRunningTime="2025-11-28 15:30:41.901905725 +0000 UTC m=+268.951697056" watchObservedRunningTime="2025-11-28 15:30:58.673281988 +0000 UTC m=+285.723073349" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.675897 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8d4l4" podStartSLOduration=35.685654941 podStartE2EDuration="2m24.675884842s" podCreationTimestamp="2025-11-28 15:28:34 +0000 UTC" firstStartedPulling="2025-11-28 15:28:36.940010132 +0000 UTC m=+143.989801443" lastFinishedPulling="2025-11-28 15:30:25.930240043 +0000 UTC m=+252.980031344" observedRunningTime="2025-11-28 15:30:41.919729967 +0000 UTC m=+268.969521278" watchObservedRunningTime="2025-11-28 15:30:58.675884842 +0000 UTC m=+285.725676163" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.676656 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=35.676649394 podStartE2EDuration="35.676649394s" podCreationTimestamp="2025-11-28 15:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:30:41.929514508 +0000 UTC m=+268.979305819" watchObservedRunningTime="2025-11-28 15:30:58.676649394 +0000 UTC m=+285.726440715" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.679511 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hm6n6","openshift-marketplace/community-operators-rgs2f","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/redhat-marketplace-tgr5n","openshift-marketplace/certified-operators-7cqfs"] Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.679631 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.683899 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.700061 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.701454 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.701429976 podStartE2EDuration="16.701429976s" podCreationTimestamp="2025-11-28 15:30:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:30:58.698307256 +0000 UTC m=+285.748098587" watchObservedRunningTime="2025-11-28 15:30:58.701429976 +0000 UTC m=+285.751221297" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.712943 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 15:30:58 crc kubenswrapper[4805]: I1128 15:30:58.984265 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.049506 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.081168 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.093645 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.147582 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.213064 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" path="/var/lib/kubelet/pods/7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa/volumes" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.214109 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" path="/var/lib/kubelet/pods/c6b04434-a30b-4e22-8c5a-9cdaca124466/volumes" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.214979 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" path="/var/lib/kubelet/pods/d53216ca-283e-4346-ace1-9fab11518bfd/volumes" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.216771 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9089841-40aa-479f-b56e-421629f24685" path="/var/lib/kubelet/pods/e9089841-40aa-479f-b56e-421629f24685/volumes" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.281799 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.297502 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.406598 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.419466 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.437906 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.469576 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.493797 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.549344 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.582148 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.787509 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.824152 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.912174 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.927184 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.945562 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 15:30:59 crc kubenswrapper[4805]: I1128 15:30:59.969768 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.007141 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.120277 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.130074 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.140212 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.140290 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.140436 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.222898 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.230270 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.284014 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.328278 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.393950 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.435388 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.626984 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.786851 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.804472 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 15:31:00 crc kubenswrapper[4805]: I1128 15:31:00.809818 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.105729 4805 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.129921 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.191124 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.194987 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.291572 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.358825 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.364265 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.414855 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.466924 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.487786 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.546660 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.635715 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.717186 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.766742 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.843346 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.850745 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 15:31:01 crc kubenswrapper[4805]: I1128 15:31:01.909772 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 15:31:02 crc kubenswrapper[4805]: I1128 15:31:02.199697 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 15:31:02 crc kubenswrapper[4805]: I1128 15:31:02.335747 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 15:31:02 crc kubenswrapper[4805]: I1128 15:31:02.345667 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 15:31:02 crc kubenswrapper[4805]: I1128 15:31:02.415466 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 15:31:02 crc kubenswrapper[4805]: I1128 15:31:02.418084 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 15:31:02 crc kubenswrapper[4805]: I1128 15:31:02.682713 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 15:31:02 crc kubenswrapper[4805]: I1128 15:31:02.824391 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 15:31:02 crc kubenswrapper[4805]: I1128 15:31:02.961470 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.137572 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.190310 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.201328 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.211173 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.228238 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.251824 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.342049 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.405099 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.480235 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.606195 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.613903 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.716084 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.840283 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 15:31:03 crc kubenswrapper[4805]: I1128 15:31:03.959410 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.021284 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.104015 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.143257 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.188123 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.337031 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.397130 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.423777 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.436748 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.440215 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.502298 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.535046 4805 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.535283 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835" gracePeriod=5 Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.583499 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.585966 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.714441 4805 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.839460 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.878773 4805 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.891112 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.920188 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 15:31:04 crc kubenswrapper[4805]: I1128 15:31:04.949113 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.224168 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.343236 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.371928 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.374972 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.379252 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.495019 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.606815 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.607058 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.634552 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.668936 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.670644 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.735214 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.798861 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.825102 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.856876 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.882447 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 15:31:05 crc kubenswrapper[4805]: I1128 15:31:05.886541 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.145671 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.171548 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.173253 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.205718 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.283739 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.352937 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.412492 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.493759 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.605256 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.675494 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.706843 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.767249 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.866909 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 15:31:06 crc kubenswrapper[4805]: I1128 15:31:06.898297 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.085004 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.176150 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.300493 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.305421 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.375115 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.424956 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.479907 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.496219 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.502493 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.505468 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.586149 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.740774 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.777768 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.803076 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.833797 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.843913 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 15:31:07 crc kubenswrapper[4805]: I1128 15:31:07.927988 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 15:31:08 crc kubenswrapper[4805]: I1128 15:31:08.100010 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 15:31:08 crc kubenswrapper[4805]: I1128 15:31:08.180521 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 15:31:08 crc kubenswrapper[4805]: I1128 15:31:08.393817 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 15:31:08 crc kubenswrapper[4805]: I1128 15:31:08.478852 4805 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 15:31:08 crc kubenswrapper[4805]: I1128 15:31:08.650806 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 15:31:08 crc kubenswrapper[4805]: I1128 15:31:08.698217 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 15:31:08 crc kubenswrapper[4805]: I1128 15:31:08.736458 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 15:31:08 crc kubenswrapper[4805]: I1128 15:31:08.782058 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 15:31:08 crc kubenswrapper[4805]: I1128 15:31:08.897994 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.004674 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.030648 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.087676 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.268977 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.394023 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.400097 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.480922 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.566694 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.570723 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.602674 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.747661 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.903485 4805 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 15:31:09 crc kubenswrapper[4805]: I1128 15:31:09.933931 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.039509 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.101167 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.101240 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.196665 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197097 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.196795 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197138 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197164 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197200 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197246 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197317 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197327 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197704 4805 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197732 4805 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197749 4805 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.197763 4805 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.206611 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.298817 4805 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.370499 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.370543 4805 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835" exitCode=137 Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.370582 4805 scope.go:117] "RemoveContainer" containerID="bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.370708 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.393046 4805 scope.go:117] "RemoveContainer" containerID="bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835" Nov 28 15:31:10 crc kubenswrapper[4805]: E1128 15:31:10.393534 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835\": container with ID starting with bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835 not found: ID does not exist" containerID="bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.393568 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835"} err="failed to get container status \"bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835\": rpc error: code = NotFound desc = could not find container \"bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835\": container with ID starting with bfedc6aef5bd09a33bf9bcd13683ec4ab1705d6dfbbbafbc80a9c59579b4e835 not found: ID does not exist" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.406508 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 15:31:10 crc kubenswrapper[4805]: I1128 15:31:10.438402 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 15:31:11 crc kubenswrapper[4805]: I1128 15:31:11.053180 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 15:31:11 crc kubenswrapper[4805]: I1128 15:31:11.213832 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 28 15:31:11 crc kubenswrapper[4805]: I1128 15:31:11.214269 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 28 15:31:11 crc kubenswrapper[4805]: I1128 15:31:11.224973 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 15:31:11 crc kubenswrapper[4805]: I1128 15:31:11.225168 4805 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="98c0ca77-e31d-46cf-a546-fb58c38be43d" Nov 28 15:31:11 crc kubenswrapper[4805]: I1128 15:31:11.229329 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 15:31:11 crc kubenswrapper[4805]: I1128 15:31:11.229401 4805 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="98c0ca77-e31d-46cf-a546-fb58c38be43d" Nov 28 15:31:11 crc kubenswrapper[4805]: I1128 15:31:11.452497 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.139631 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dhzqc"] Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.140330 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dhzqc" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerName="registry-server" containerID="cri-o://eb15017ccdf744b374fcdaec4b8aea1b2514300d65bf7b400c26f17a7a53f4fd" gracePeriod=30 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.151603 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8d4l4"] Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.152229 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8d4l4" podUID="db028335-3e13-44c6-844f-318168230b30" containerName="registry-server" containerID="cri-o://fb2dcff113a351d2085a3e177008f9bcc2129e6e43d6b47f4219a16fd74664ae" gracePeriod=30 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.185113 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dzb6r"] Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.185531 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" podUID="ed210d08-27a0-4b04-aa18-1387ca89d44a" containerName="marketplace-operator" containerID="cri-o://783b3213f3e536db2bb1a0188bf5cc31bd1ebe24c4a18274ee10234a7c598baa" gracePeriod=30 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.212422 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2fzf"] Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.212464 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s6fmm"] Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.212663 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s6fmm" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerName="registry-server" containerID="cri-o://c567e5a631e2408cceca537997282c14a19fe198f99f5c050e2fbce23b0adef2" gracePeriod=30 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.212782 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s2fzf" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerName="registry-server" containerID="cri-o://644cd402bc258a78446db53307622ef06b0c91d2e4bc740774eef3123ecb4ddc" gracePeriod=30 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223347 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ld8lt"] Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223668 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerName="registry-server" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223692 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerName="registry-server" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223708 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerName="extract-utilities" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223717 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerName="extract-utilities" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223728 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223736 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223748 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerName="extract-content" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223756 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerName="extract-content" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223766 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" containerName="extract-utilities" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223774 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" containerName="extract-utilities" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223790 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" containerName="extract-content" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223798 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" containerName="extract-content" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223808 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerName="extract-utilities" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223815 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerName="extract-utilities" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223827 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerName="extract-content" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223834 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerName="extract-content" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223849 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72212593-ffb8-48a1-a673-5532833d9134" containerName="installer" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223856 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="72212593-ffb8-48a1-a673-5532833d9134" containerName="installer" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223867 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerName="registry-server" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223874 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerName="registry-server" Nov 28 15:31:19 crc kubenswrapper[4805]: E1128 15:31:19.223887 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" containerName="registry-server" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.223895 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" containerName="registry-server" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.224024 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="72212593-ffb8-48a1-a673-5532833d9134" containerName="installer" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.224039 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6b04434-a30b-4e22-8c5a-9cdaca124466" containerName="registry-server" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.224052 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d53216ca-283e-4346-ace1-9fab11518bfd" containerName="registry-server" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.224060 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.224072 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbb256b-2735-4b8b-91ed-cd1ee9edd6fa" containerName="registry-server" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.224530 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.227094 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ld8lt"] Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.416143 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mshxs\" (UniqueName: \"kubernetes.io/projected/82e5203a-95bd-4916-8cec-43da144a6c0a-kube-api-access-mshxs\") pod \"marketplace-operator-79b997595-ld8lt\" (UID: \"82e5203a-95bd-4916-8cec-43da144a6c0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.416205 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/82e5203a-95bd-4916-8cec-43da144a6c0a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ld8lt\" (UID: \"82e5203a-95bd-4916-8cec-43da144a6c0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.416247 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/82e5203a-95bd-4916-8cec-43da144a6c0a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ld8lt\" (UID: \"82e5203a-95bd-4916-8cec-43da144a6c0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.439520 4805 generic.go:334] "Generic (PLEG): container finished" podID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerID="644cd402bc258a78446db53307622ef06b0c91d2e4bc740774eef3123ecb4ddc" exitCode=0 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.439627 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2fzf" event={"ID":"eff8293e-510a-4cf5-afe2-5fefc4da0d3f","Type":"ContainerDied","Data":"644cd402bc258a78446db53307622ef06b0c91d2e4bc740774eef3123ecb4ddc"} Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.442897 4805 generic.go:334] "Generic (PLEG): container finished" podID="db028335-3e13-44c6-844f-318168230b30" containerID="fb2dcff113a351d2085a3e177008f9bcc2129e6e43d6b47f4219a16fd74664ae" exitCode=0 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.442978 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d4l4" event={"ID":"db028335-3e13-44c6-844f-318168230b30","Type":"ContainerDied","Data":"fb2dcff113a351d2085a3e177008f9bcc2129e6e43d6b47f4219a16fd74664ae"} Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.445500 4805 generic.go:334] "Generic (PLEG): container finished" podID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerID="eb15017ccdf744b374fcdaec4b8aea1b2514300d65bf7b400c26f17a7a53f4fd" exitCode=0 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.445614 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhzqc" event={"ID":"bf218599-d9d7-4e0e-a2ab-0949ded9fa58","Type":"ContainerDied","Data":"eb15017ccdf744b374fcdaec4b8aea1b2514300d65bf7b400c26f17a7a53f4fd"} Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.447984 4805 generic.go:334] "Generic (PLEG): container finished" podID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerID="c567e5a631e2408cceca537997282c14a19fe198f99f5c050e2fbce23b0adef2" exitCode=0 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.448062 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6fmm" event={"ID":"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07","Type":"ContainerDied","Data":"c567e5a631e2408cceca537997282c14a19fe198f99f5c050e2fbce23b0adef2"} Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.449882 4805 generic.go:334] "Generic (PLEG): container finished" podID="ed210d08-27a0-4b04-aa18-1387ca89d44a" containerID="783b3213f3e536db2bb1a0188bf5cc31bd1ebe24c4a18274ee10234a7c598baa" exitCode=0 Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.449942 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" event={"ID":"ed210d08-27a0-4b04-aa18-1387ca89d44a","Type":"ContainerDied","Data":"783b3213f3e536db2bb1a0188bf5cc31bd1ebe24c4a18274ee10234a7c598baa"} Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.517164 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mshxs\" (UniqueName: \"kubernetes.io/projected/82e5203a-95bd-4916-8cec-43da144a6c0a-kube-api-access-mshxs\") pod \"marketplace-operator-79b997595-ld8lt\" (UID: \"82e5203a-95bd-4916-8cec-43da144a6c0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.517207 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/82e5203a-95bd-4916-8cec-43da144a6c0a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ld8lt\" (UID: \"82e5203a-95bd-4916-8cec-43da144a6c0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.517233 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/82e5203a-95bd-4916-8cec-43da144a6c0a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ld8lt\" (UID: \"82e5203a-95bd-4916-8cec-43da144a6c0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.519420 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/82e5203a-95bd-4916-8cec-43da144a6c0a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ld8lt\" (UID: \"82e5203a-95bd-4916-8cec-43da144a6c0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.523428 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/82e5203a-95bd-4916-8cec-43da144a6c0a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ld8lt\" (UID: \"82e5203a-95bd-4916-8cec-43da144a6c0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.544824 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mshxs\" (UniqueName: \"kubernetes.io/projected/82e5203a-95bd-4916-8cec-43da144a6c0a-kube-api-access-mshxs\") pod \"marketplace-operator-79b997595-ld8lt\" (UID: \"82e5203a-95bd-4916-8cec-43da144a6c0a\") " pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.595574 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.599348 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.605921 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.606745 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.610368 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.640593 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.725391 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-catalog-content\") pod \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.725768 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phqdp\" (UniqueName: \"kubernetes.io/projected/db028335-3e13-44c6-844f-318168230b30-kube-api-access-phqdp\") pod \"db028335-3e13-44c6-844f-318168230b30\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.725833 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-catalog-content\") pod \"db028335-3e13-44c6-844f-318168230b30\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.725874 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lcm5\" (UniqueName: \"kubernetes.io/projected/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-kube-api-access-9lcm5\") pod \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.725898 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gssxt\" (UniqueName: \"kubernetes.io/projected/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-kube-api-access-gssxt\") pod \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.725935 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-trusted-ca\") pod \"ed210d08-27a0-4b04-aa18-1387ca89d44a\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.725958 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-operator-metrics\") pod \"ed210d08-27a0-4b04-aa18-1387ca89d44a\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.725981 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8pmv\" (UniqueName: \"kubernetes.io/projected/ed210d08-27a0-4b04-aa18-1387ca89d44a-kube-api-access-d8pmv\") pod \"ed210d08-27a0-4b04-aa18-1387ca89d44a\" (UID: \"ed210d08-27a0-4b04-aa18-1387ca89d44a\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.726013 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-catalog-content\") pod \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.726033 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-utilities\") pod \"db028335-3e13-44c6-844f-318168230b30\" (UID: \"db028335-3e13-44c6-844f-318168230b30\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.726048 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-utilities\") pod \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\" (UID: \"bf218599-d9d7-4e0e-a2ab-0949ded9fa58\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.726068 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-utilities\") pod \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\" (UID: \"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.729077 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ed210d08-27a0-4b04-aa18-1387ca89d44a" (UID: "ed210d08-27a0-4b04-aa18-1387ca89d44a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.730068 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-utilities" (OuterVolumeSpecName: "utilities") pod "bf218599-d9d7-4e0e-a2ab-0949ded9fa58" (UID: "bf218599-d9d7-4e0e-a2ab-0949ded9fa58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.733751 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-utilities" (OuterVolumeSpecName: "utilities") pod "bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" (UID: "bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.734851 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-utilities" (OuterVolumeSpecName: "utilities") pod "db028335-3e13-44c6-844f-318168230b30" (UID: "db028335-3e13-44c6-844f-318168230b30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.734983 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-kube-api-access-gssxt" (OuterVolumeSpecName: "kube-api-access-gssxt") pod "bf218599-d9d7-4e0e-a2ab-0949ded9fa58" (UID: "bf218599-d9d7-4e0e-a2ab-0949ded9fa58"). InnerVolumeSpecName "kube-api-access-gssxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.748614 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-kube-api-access-9lcm5" (OuterVolumeSpecName: "kube-api-access-9lcm5") pod "bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" (UID: "bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07"). InnerVolumeSpecName "kube-api-access-9lcm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.748635 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ed210d08-27a0-4b04-aa18-1387ca89d44a" (UID: "ed210d08-27a0-4b04-aa18-1387ca89d44a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.759573 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db028335-3e13-44c6-844f-318168230b30-kube-api-access-phqdp" (OuterVolumeSpecName: "kube-api-access-phqdp") pod "db028335-3e13-44c6-844f-318168230b30" (UID: "db028335-3e13-44c6-844f-318168230b30"). InnerVolumeSpecName "kube-api-access-phqdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.761896 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed210d08-27a0-4b04-aa18-1387ca89d44a-kube-api-access-d8pmv" (OuterVolumeSpecName: "kube-api-access-d8pmv") pod "ed210d08-27a0-4b04-aa18-1387ca89d44a" (UID: "ed210d08-27a0-4b04-aa18-1387ca89d44a"). InnerVolumeSpecName "kube-api-access-d8pmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.794380 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf218599-d9d7-4e0e-a2ab-0949ded9fa58" (UID: "bf218599-d9d7-4e0e-a2ab-0949ded9fa58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.814684 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db028335-3e13-44c6-844f-318168230b30" (UID: "db028335-3e13-44c6-844f-318168230b30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.827903 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-catalog-content\") pod \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.827967 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-utilities\") pod \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828014 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlkbz\" (UniqueName: \"kubernetes.io/projected/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-kube-api-access-wlkbz\") pod \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\" (UID: \"eff8293e-510a-4cf5-afe2-5fefc4da0d3f\") " Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828225 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828244 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lcm5\" (UniqueName: \"kubernetes.io/projected/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-kube-api-access-9lcm5\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828256 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gssxt\" (UniqueName: \"kubernetes.io/projected/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-kube-api-access-gssxt\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828265 4805 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828275 4805 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed210d08-27a0-4b04-aa18-1387ca89d44a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828286 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8pmv\" (UniqueName: \"kubernetes.io/projected/ed210d08-27a0-4b04-aa18-1387ca89d44a-kube-api-access-d8pmv\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828293 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db028335-3e13-44c6-844f-318168230b30-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828302 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828310 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828318 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf218599-d9d7-4e0e-a2ab-0949ded9fa58-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.828327 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phqdp\" (UniqueName: \"kubernetes.io/projected/db028335-3e13-44c6-844f-318168230b30-kube-api-access-phqdp\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.829804 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-utilities" (OuterVolumeSpecName: "utilities") pod "eff8293e-510a-4cf5-afe2-5fefc4da0d3f" (UID: "eff8293e-510a-4cf5-afe2-5fefc4da0d3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.831419 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-kube-api-access-wlkbz" (OuterVolumeSpecName: "kube-api-access-wlkbz") pod "eff8293e-510a-4cf5-afe2-5fefc4da0d3f" (UID: "eff8293e-510a-4cf5-afe2-5fefc4da0d3f"). InnerVolumeSpecName "kube-api-access-wlkbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.843547 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ld8lt"] Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.846352 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eff8293e-510a-4cf5-afe2-5fefc4da0d3f" (UID: "eff8293e-510a-4cf5-afe2-5fefc4da0d3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.853506 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" (UID: "bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.929323 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.929367 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.929379 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:19 crc kubenswrapper[4805]: I1128 15:31:19.929389 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlkbz\" (UniqueName: \"kubernetes.io/projected/eff8293e-510a-4cf5-afe2-5fefc4da0d3f-kube-api-access-wlkbz\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.457727 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6fmm" event={"ID":"bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07","Type":"ContainerDied","Data":"5a827dd1ed486794b97b8956f8e2aeaf45cf10b6a237e6d2dfd792bd559ace05"} Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.457746 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s6fmm" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.458438 4805 scope.go:117] "RemoveContainer" containerID="c567e5a631e2408cceca537997282c14a19fe198f99f5c050e2fbce23b0adef2" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.459526 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" event={"ID":"82e5203a-95bd-4916-8cec-43da144a6c0a","Type":"ContainerStarted","Data":"d71426adeae93670ad2d8a3900d2d40ee12b73102598ba7054b8abc69be38595"} Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.460509 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.460530 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" event={"ID":"82e5203a-95bd-4916-8cec-43da144a6c0a","Type":"ContainerStarted","Data":"d63603ed893bc7eb440dacc9877b5874c73f8d6ff6b01c798edd776b620713e0"} Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.463388 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" event={"ID":"ed210d08-27a0-4b04-aa18-1387ca89d44a","Type":"ContainerDied","Data":"993142643a4512e108f050c2d339d7c8cc8661e3ec5b3dbf87a45ccd4bb30511"} Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.463406 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dzb6r" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.465877 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.466377 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2fzf" event={"ID":"eff8293e-510a-4cf5-afe2-5fefc4da0d3f","Type":"ContainerDied","Data":"3afba656b135a59d75e8f19b1543a8b4738661d9bc002869e66af7a4c486779b"} Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.466403 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2fzf" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.476697 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d4l4" event={"ID":"db028335-3e13-44c6-844f-318168230b30","Type":"ContainerDied","Data":"e40e55c25a4094af294abe198a08c4afe5a2bd3f4829dc3c3a70182d263102d4"} Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.476749 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8d4l4" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.480697 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhzqc" event={"ID":"bf218599-d9d7-4e0e-a2ab-0949ded9fa58","Type":"ContainerDied","Data":"71cc22f73e7b72ee7fcc570e09b6438c0cfdab3975a9f0d0268b6d4731971286"} Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.480786 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dhzqc" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.484979 4805 scope.go:117] "RemoveContainer" containerID="82935b7ba460ce053a7e939956fdb891a539aa81a626fc7085c132043faf3c6c" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.487748 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ld8lt" podStartSLOduration=1.487728344 podStartE2EDuration="1.487728344s" podCreationTimestamp="2025-11-28 15:31:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:31:20.479195929 +0000 UTC m=+307.528987250" watchObservedRunningTime="2025-11-28 15:31:20.487728344 +0000 UTC m=+307.537519665" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.503433 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dzb6r"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.516990 4805 scope.go:117] "RemoveContainer" containerID="47d629bf37f80023c3e1c47273a4700feac831b3d8c5c56216e89bb4a67a0406" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.524548 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dzb6r"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.554084 4805 scope.go:117] "RemoveContainer" containerID="783b3213f3e536db2bb1a0188bf5cc31bd1ebe24c4a18274ee10234a7c598baa" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.604405 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s6fmm"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.605094 4805 scope.go:117] "RemoveContainer" containerID="644cd402bc258a78446db53307622ef06b0c91d2e4bc740774eef3123ecb4ddc" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.614907 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s6fmm"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.627615 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8d4l4"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.640135 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8d4l4"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.650743 4805 scope.go:117] "RemoveContainer" containerID="f0fe943fd27b3d6f97690a00b062e43db356f7d7b8cb739b4c66f86f88d891dc" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.668934 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dhzqc"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.672424 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dhzqc"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.676174 4805 scope.go:117] "RemoveContainer" containerID="aa867351a96ffb824c7c8c4ee504065aeecb3924e0437e115806297ce9d380fa" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.679173 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2fzf"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.685512 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2fzf"] Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.691568 4805 scope.go:117] "RemoveContainer" containerID="fb2dcff113a351d2085a3e177008f9bcc2129e6e43d6b47f4219a16fd74664ae" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.706829 4805 scope.go:117] "RemoveContainer" containerID="d3bb12c3972d097f73396516c8a9264bccff569673db36f991f940f635c24120" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.730685 4805 scope.go:117] "RemoveContainer" containerID="5bb049dff1328fb61ad59f6d28dd769a3323dc3facceb6099f58a0130398da4e" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.744297 4805 scope.go:117] "RemoveContainer" containerID="eb15017ccdf744b374fcdaec4b8aea1b2514300d65bf7b400c26f17a7a53f4fd" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.755959 4805 scope.go:117] "RemoveContainer" containerID="972b766ed63a8c52bce906d61dfb675dbcb740b2de63e5b42c513d4ab09b8f25" Nov 28 15:31:20 crc kubenswrapper[4805]: I1128 15:31:20.770997 4805 scope.go:117] "RemoveContainer" containerID="8b142cd70a956ee827eb8f17fa59358d1f273ab2ef0f7d0adecded046621aae4" Nov 28 15:31:21 crc kubenswrapper[4805]: I1128 15:31:21.210993 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" path="/var/lib/kubelet/pods/bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07/volumes" Nov 28 15:31:21 crc kubenswrapper[4805]: I1128 15:31:21.211753 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" path="/var/lib/kubelet/pods/bf218599-d9d7-4e0e-a2ab-0949ded9fa58/volumes" Nov 28 15:31:21 crc kubenswrapper[4805]: I1128 15:31:21.212474 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db028335-3e13-44c6-844f-318168230b30" path="/var/lib/kubelet/pods/db028335-3e13-44c6-844f-318168230b30/volumes" Nov 28 15:31:21 crc kubenswrapper[4805]: I1128 15:31:21.213646 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed210d08-27a0-4b04-aa18-1387ca89d44a" path="/var/lib/kubelet/pods/ed210d08-27a0-4b04-aa18-1387ca89d44a/volumes" Nov 28 15:31:21 crc kubenswrapper[4805]: I1128 15:31:21.214070 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" path="/var/lib/kubelet/pods/eff8293e-510a-4cf5-afe2-5fefc4da0d3f/volumes" Nov 28 15:32:05 crc kubenswrapper[4805]: I1128 15:32:05.860106 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q6czs"] Nov 28 15:32:05 crc kubenswrapper[4805]: I1128 15:32:05.863465 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" podUID="85065f7f-4262-49e8-8f55-4df7edc2b8a7" containerName="controller-manager" containerID="cri-o://41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7" gracePeriod=30 Nov 28 15:32:05 crc kubenswrapper[4805]: I1128 15:32:05.962926 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x"] Nov 28 15:32:05 crc kubenswrapper[4805]: I1128 15:32:05.963493 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" podUID="f9430f70-1c4a-47af-813f-76079af84e5e" containerName="route-controller-manager" containerID="cri-o://4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6" gracePeriod=30 Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.194256 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.268031 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.325556 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-proxy-ca-bundles\") pod \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.325660 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnrsh\" (UniqueName: \"kubernetes.io/projected/85065f7f-4262-49e8-8f55-4df7edc2b8a7-kube-api-access-vnrsh\") pod \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.325728 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-client-ca\") pod \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.325750 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85065f7f-4262-49e8-8f55-4df7edc2b8a7-serving-cert\") pod \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.325774 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-config\") pod \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\" (UID: \"85065f7f-4262-49e8-8f55-4df7edc2b8a7\") " Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.326701 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "85065f7f-4262-49e8-8f55-4df7edc2b8a7" (UID: "85065f7f-4262-49e8-8f55-4df7edc2b8a7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.326768 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-client-ca" (OuterVolumeSpecName: "client-ca") pod "85065f7f-4262-49e8-8f55-4df7edc2b8a7" (UID: "85065f7f-4262-49e8-8f55-4df7edc2b8a7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.326859 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-config" (OuterVolumeSpecName: "config") pod "85065f7f-4262-49e8-8f55-4df7edc2b8a7" (UID: "85065f7f-4262-49e8-8f55-4df7edc2b8a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.331174 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85065f7f-4262-49e8-8f55-4df7edc2b8a7-kube-api-access-vnrsh" (OuterVolumeSpecName: "kube-api-access-vnrsh") pod "85065f7f-4262-49e8-8f55-4df7edc2b8a7" (UID: "85065f7f-4262-49e8-8f55-4df7edc2b8a7"). InnerVolumeSpecName "kube-api-access-vnrsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.331628 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85065f7f-4262-49e8-8f55-4df7edc2b8a7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "85065f7f-4262-49e8-8f55-4df7edc2b8a7" (UID: "85065f7f-4262-49e8-8f55-4df7edc2b8a7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.426817 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9430f70-1c4a-47af-813f-76079af84e5e-serving-cert\") pod \"f9430f70-1c4a-47af-813f-76079af84e5e\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.426890 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-config\") pod \"f9430f70-1c4a-47af-813f-76079af84e5e\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.427347 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfrtd\" (UniqueName: \"kubernetes.io/projected/f9430f70-1c4a-47af-813f-76079af84e5e-kube-api-access-sfrtd\") pod \"f9430f70-1c4a-47af-813f-76079af84e5e\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.428097 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-client-ca\") pod \"f9430f70-1c4a-47af-813f-76079af84e5e\" (UID: \"f9430f70-1c4a-47af-813f-76079af84e5e\") " Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.428389 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnrsh\" (UniqueName: \"kubernetes.io/projected/85065f7f-4262-49e8-8f55-4df7edc2b8a7-kube-api-access-vnrsh\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.428412 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.428425 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85065f7f-4262-49e8-8f55-4df7edc2b8a7-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.428437 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.428447 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/85065f7f-4262-49e8-8f55-4df7edc2b8a7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.428898 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-config" (OuterVolumeSpecName: "config") pod "f9430f70-1c4a-47af-813f-76079af84e5e" (UID: "f9430f70-1c4a-47af-813f-76079af84e5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.429028 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-client-ca" (OuterVolumeSpecName: "client-ca") pod "f9430f70-1c4a-47af-813f-76079af84e5e" (UID: "f9430f70-1c4a-47af-813f-76079af84e5e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.431082 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9430f70-1c4a-47af-813f-76079af84e5e-kube-api-access-sfrtd" (OuterVolumeSpecName: "kube-api-access-sfrtd") pod "f9430f70-1c4a-47af-813f-76079af84e5e" (UID: "f9430f70-1c4a-47af-813f-76079af84e5e"). InnerVolumeSpecName "kube-api-access-sfrtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.432085 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9430f70-1c4a-47af-813f-76079af84e5e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f9430f70-1c4a-47af-813f-76079af84e5e" (UID: "f9430f70-1c4a-47af-813f-76079af84e5e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.529560 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.529622 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9430f70-1c4a-47af-813f-76079af84e5e-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.529649 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9430f70-1c4a-47af-813f-76079af84e5e-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.529676 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfrtd\" (UniqueName: \"kubernetes.io/projected/f9430f70-1c4a-47af-813f-76079af84e5e-kube-api-access-sfrtd\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.775160 4805 generic.go:334] "Generic (PLEG): container finished" podID="f9430f70-1c4a-47af-813f-76079af84e5e" containerID="4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6" exitCode=0 Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.775388 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" event={"ID":"f9430f70-1c4a-47af-813f-76079af84e5e","Type":"ContainerDied","Data":"4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6"} Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.775539 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.775433 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x" event={"ID":"f9430f70-1c4a-47af-813f-76079af84e5e","Type":"ContainerDied","Data":"d5a14efd4379f6d18bcea91e7410e861c9fda8fe6edea508196636661ffb4719"} Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.776114 4805 scope.go:117] "RemoveContainer" containerID="4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.778797 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.778709 4805 generic.go:334] "Generic (PLEG): container finished" podID="85065f7f-4262-49e8-8f55-4df7edc2b8a7" containerID="41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7" exitCode=0 Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.778887 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" event={"ID":"85065f7f-4262-49e8-8f55-4df7edc2b8a7","Type":"ContainerDied","Data":"41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7"} Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.778952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q6czs" event={"ID":"85065f7f-4262-49e8-8f55-4df7edc2b8a7","Type":"ContainerDied","Data":"7143883eed036ca11fddf45a52e706f6f64cb834bbb9b760fbace5a791a45bea"} Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.815071 4805 scope.go:117] "RemoveContainer" containerID="4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6" Nov 28 15:32:06 crc kubenswrapper[4805]: E1128 15:32:06.815711 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6\": container with ID starting with 4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6 not found: ID does not exist" containerID="4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.815787 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6"} err="failed to get container status \"4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6\": rpc error: code = NotFound desc = could not find container \"4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6\": container with ID starting with 4278771aee039b9aa373e847f868c5f0e48e7d6ebbc060062b44409d6a8b52d6 not found: ID does not exist" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.815837 4805 scope.go:117] "RemoveContainer" containerID="41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.826887 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x"] Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.840201 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dcr6x"] Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.844632 4805 scope.go:117] "RemoveContainer" containerID="41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7" Nov 28 15:32:06 crc kubenswrapper[4805]: E1128 15:32:06.845298 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7\": container with ID starting with 41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7 not found: ID does not exist" containerID="41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.845346 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7"} err="failed to get container status \"41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7\": rpc error: code = NotFound desc = could not find container \"41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7\": container with ID starting with 41086feb83f82c778849aa20125424b3bd6a85d7bafe1a555e528788a9303da7 not found: ID does not exist" Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.847108 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q6czs"] Nov 28 15:32:06 crc kubenswrapper[4805]: I1128 15:32:06.853020 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q6czs"] Nov 28 15:32:06 crc kubenswrapper[4805]: E1128 15:32:06.908026 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85065f7f_4262_49e8_8f55_4df7edc2b8a7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9430f70_1c4a_47af_813f_76079af84e5e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9430f70_1c4a_47af_813f_76079af84e5e.slice/crio-d5a14efd4379f6d18bcea91e7410e861c9fda8fe6edea508196636661ffb4719\": RecentStats: unable to find data in memory cache]" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.024704 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr"] Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025010 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025034 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025052 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db028335-3e13-44c6-844f-318168230b30" containerName="extract-utilities" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025062 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="db028335-3e13-44c6-844f-318168230b30" containerName="extract-utilities" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025077 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerName="extract-utilities" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025090 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerName="extract-utilities" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025106 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db028335-3e13-44c6-844f-318168230b30" containerName="extract-content" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025117 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="db028335-3e13-44c6-844f-318168230b30" containerName="extract-content" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025129 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerName="extract-utilities" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025140 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerName="extract-utilities" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025158 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9430f70-1c4a-47af-813f-76079af84e5e" containerName="route-controller-manager" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025169 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9430f70-1c4a-47af-813f-76079af84e5e" containerName="route-controller-manager" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025182 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerName="extract-content" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025192 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerName="extract-content" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025209 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025219 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025237 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed210d08-27a0-4b04-aa18-1387ca89d44a" containerName="marketplace-operator" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025250 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed210d08-27a0-4b04-aa18-1387ca89d44a" containerName="marketplace-operator" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025266 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025275 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025293 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerName="extract-content" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025303 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerName="extract-content" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025316 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db028335-3e13-44c6-844f-318168230b30" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025325 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="db028335-3e13-44c6-844f-318168230b30" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025337 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerName="extract-utilities" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025347 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerName="extract-utilities" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025383 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85065f7f-4262-49e8-8f55-4df7edc2b8a7" containerName="controller-manager" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025395 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="85065f7f-4262-49e8-8f55-4df7edc2b8a7" containerName="controller-manager" Nov 28 15:32:07 crc kubenswrapper[4805]: E1128 15:32:07.025407 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerName="extract-content" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025418 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerName="extract-content" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025566 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf218599-d9d7-4e0e-a2ab-0949ded9fa58" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025584 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff8293e-510a-4cf5-afe2-5fefc4da0d3f" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025597 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="85065f7f-4262-49e8-8f55-4df7edc2b8a7" containerName="controller-manager" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025617 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="db028335-3e13-44c6-844f-318168230b30" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025630 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33bc5f-4d48-4cc5-899e-cd5d0d5a4a07" containerName="registry-server" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025647 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9430f70-1c4a-47af-813f-76079af84e5e" containerName="route-controller-manager" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.025658 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed210d08-27a0-4b04-aa18-1387ca89d44a" containerName="marketplace-operator" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.026185 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.029751 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.029938 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.030083 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.030261 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.030465 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.034310 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-79794f6f4c-q65v6"] Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.035082 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.035466 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-config\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.035614 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-client-ca\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.035717 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2b6j\" (UniqueName: \"kubernetes.io/projected/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-kube-api-access-v2b6j\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.035841 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-serving-cert\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.041051 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.044451 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.049081 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.049479 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.049968 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.050804 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.050875 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.050975 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr"] Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.054727 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79794f6f4c-q65v6"] Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.055497 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.136905 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-config\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.136959 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-client-ca\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.137000 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2b6j\" (UniqueName: \"kubernetes.io/projected/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-kube-api-access-v2b6j\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.137046 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-serving-cert\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.138323 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-config\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.138331 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-client-ca\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.142587 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-serving-cert\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.160451 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2b6j\" (UniqueName: \"kubernetes.io/projected/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-kube-api-access-v2b6j\") pod \"route-controller-manager-58c85c56bf-pzjjr\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.213182 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85065f7f-4262-49e8-8f55-4df7edc2b8a7" path="/var/lib/kubelet/pods/85065f7f-4262-49e8-8f55-4df7edc2b8a7/volumes" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.214315 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9430f70-1c4a-47af-813f-76079af84e5e" path="/var/lib/kubelet/pods/f9430f70-1c4a-47af-813f-76079af84e5e/volumes" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.238449 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b35f118-5646-441c-a833-ba6fb427dfdb-serving-cert\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.238509 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p82hl\" (UniqueName: \"kubernetes.io/projected/5b35f118-5646-441c-a833-ba6fb427dfdb-kube-api-access-p82hl\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.238681 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-proxy-ca-bundles\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.238865 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-config\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.239025 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-client-ca\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.339931 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-client-ca\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.340016 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b35f118-5646-441c-a833-ba6fb427dfdb-serving-cert\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.340053 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p82hl\" (UniqueName: \"kubernetes.io/projected/5b35f118-5646-441c-a833-ba6fb427dfdb-kube-api-access-p82hl\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.340112 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-proxy-ca-bundles\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.340161 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-config\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.341348 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-client-ca\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.342057 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-config\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.343336 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-proxy-ca-bundles\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.346404 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b35f118-5646-441c-a833-ba6fb427dfdb-serving-cert\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.357741 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.359995 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p82hl\" (UniqueName: \"kubernetes.io/projected/5b35f118-5646-441c-a833-ba6fb427dfdb-kube-api-access-p82hl\") pod \"controller-manager-79794f6f4c-q65v6\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.371459 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.422811 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-79794f6f4c-q65v6"] Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.439403 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr"] Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.640319 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-79794f6f4c-q65v6"] Nov 28 15:32:07 crc kubenswrapper[4805]: W1128 15:32:07.659968 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b35f118_5646_441c_a833_ba6fb427dfdb.slice/crio-6efc8768774bef1653ab860aa16067a73940ef8d4e98ae60462e9bda84195088 WatchSource:0}: Error finding container 6efc8768774bef1653ab860aa16067a73940ef8d4e98ae60462e9bda84195088: Status 404 returned error can't find the container with id 6efc8768774bef1653ab860aa16067a73940ef8d4e98ae60462e9bda84195088 Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.696790 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr"] Nov 28 15:32:07 crc kubenswrapper[4805]: W1128 15:32:07.700642 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08c89ff9_7f27_4bd2_8d93_ec70804fbfed.slice/crio-8b4b083a5be887f825ec6df59d2cb5e6244fbffd9e951cd961fda009c10c92d9 WatchSource:0}: Error finding container 8b4b083a5be887f825ec6df59d2cb5e6244fbffd9e951cd961fda009c10c92d9: Status 404 returned error can't find the container with id 8b4b083a5be887f825ec6df59d2cb5e6244fbffd9e951cd961fda009c10c92d9 Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.784949 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" event={"ID":"5b35f118-5646-441c-a833-ba6fb427dfdb","Type":"ContainerStarted","Data":"6efc8768774bef1653ab860aa16067a73940ef8d4e98ae60462e9bda84195088"} Nov 28 15:32:07 crc kubenswrapper[4805]: I1128 15:32:07.785878 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" event={"ID":"08c89ff9-7f27-4bd2-8d93-ec70804fbfed","Type":"ContainerStarted","Data":"8b4b083a5be887f825ec6df59d2cb5e6244fbffd9e951cd961fda009c10c92d9"} Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.792878 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" event={"ID":"5b35f118-5646-441c-a833-ba6fb427dfdb","Type":"ContainerStarted","Data":"db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e"} Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.793463 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.793020 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" podUID="5b35f118-5646-441c-a833-ba6fb427dfdb" containerName="controller-manager" containerID="cri-o://db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e" gracePeriod=30 Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.794955 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" event={"ID":"08c89ff9-7f27-4bd2-8d93-ec70804fbfed","Type":"ContainerStarted","Data":"e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d"} Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.795069 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" podUID="08c89ff9-7f27-4bd2-8d93-ec70804fbfed" containerName="route-controller-manager" containerID="cri-o://e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d" gracePeriod=30 Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.795162 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.799348 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.799846 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.814673 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" podStartSLOduration=3.814653215 podStartE2EDuration="3.814653215s" podCreationTimestamp="2025-11-28 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:32:08.814034007 +0000 UTC m=+355.863825318" watchObservedRunningTime="2025-11-28 15:32:08.814653215 +0000 UTC m=+355.864444526" Nov 28 15:32:08 crc kubenswrapper[4805]: I1128 15:32:08.861585 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" podStartSLOduration=3.8615640620000002 podStartE2EDuration="3.861564062s" podCreationTimestamp="2025-11-28 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:32:08.857972699 +0000 UTC m=+355.907764030" watchObservedRunningTime="2025-11-28 15:32:08.861564062 +0000 UTC m=+355.911355373" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.144742 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.148731 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.172343 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4"] Nov 28 15:32:09 crc kubenswrapper[4805]: E1128 15:32:09.172580 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b35f118-5646-441c-a833-ba6fb427dfdb" containerName="controller-manager" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.172595 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b35f118-5646-441c-a833-ba6fb427dfdb" containerName="controller-manager" Nov 28 15:32:09 crc kubenswrapper[4805]: E1128 15:32:09.172607 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c89ff9-7f27-4bd2-8d93-ec70804fbfed" containerName="route-controller-manager" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.172615 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c89ff9-7f27-4bd2-8d93-ec70804fbfed" containerName="route-controller-manager" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.172708 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b35f118-5646-441c-a833-ba6fb427dfdb" containerName="controller-manager" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.172717 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="08c89ff9-7f27-4bd2-8d93-ec70804fbfed" containerName="route-controller-manager" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.173037 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.181501 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4"] Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.267608 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-config\") pod \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.267737 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p82hl\" (UniqueName: \"kubernetes.io/projected/5b35f118-5646-441c-a833-ba6fb427dfdb-kube-api-access-p82hl\") pod \"5b35f118-5646-441c-a833-ba6fb427dfdb\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.268108 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b35f118-5646-441c-a833-ba6fb427dfdb-serving-cert\") pod \"5b35f118-5646-441c-a833-ba6fb427dfdb\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.268210 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-client-ca\") pod \"5b35f118-5646-441c-a833-ba6fb427dfdb\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.268315 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-config\") pod \"5b35f118-5646-441c-a833-ba6fb427dfdb\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.268410 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-serving-cert\") pod \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.268488 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-client-ca\") pod \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.268564 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2b6j\" (UniqueName: \"kubernetes.io/projected/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-kube-api-access-v2b6j\") pod \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\" (UID: \"08c89ff9-7f27-4bd2-8d93-ec70804fbfed\") " Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269050 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-proxy-ca-bundles\") pod \"5b35f118-5646-441c-a833-ba6fb427dfdb\" (UID: \"5b35f118-5646-441c-a833-ba6fb427dfdb\") " Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.268838 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-client-ca" (OuterVolumeSpecName: "client-ca") pod "5b35f118-5646-441c-a833-ba6fb427dfdb" (UID: "5b35f118-5646-441c-a833-ba6fb427dfdb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269045 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-config" (OuterVolumeSpecName: "config") pod "08c89ff9-7f27-4bd2-8d93-ec70804fbfed" (UID: "08c89ff9-7f27-4bd2-8d93-ec70804fbfed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269124 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-config" (OuterVolumeSpecName: "config") pod "5b35f118-5646-441c-a833-ba6fb427dfdb" (UID: "5b35f118-5646-441c-a833-ba6fb427dfdb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269118 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-client-ca" (OuterVolumeSpecName: "client-ca") pod "08c89ff9-7f27-4bd2-8d93-ec70804fbfed" (UID: "08c89ff9-7f27-4bd2-8d93-ec70804fbfed"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269386 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf109684-739e-4bdd-94bb-8ba8d975b187-serving-cert\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269556 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-client-ca\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269649 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-config\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269702 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5b35f118-5646-441c-a833-ba6fb427dfdb" (UID: "5b35f118-5646-441c-a833-ba6fb427dfdb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269784 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8djtk\" (UniqueName: \"kubernetes.io/projected/bf109684-739e-4bdd-94bb-8ba8d975b187-kube-api-access-8djtk\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.269934 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-proxy-ca-bundles\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.270136 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.270202 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.270271 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.270344 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.270427 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5b35f118-5646-441c-a833-ba6fb427dfdb-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.273514 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "08c89ff9-7f27-4bd2-8d93-ec70804fbfed" (UID: "08c89ff9-7f27-4bd2-8d93-ec70804fbfed"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.273526 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-kube-api-access-v2b6j" (OuterVolumeSpecName: "kube-api-access-v2b6j") pod "08c89ff9-7f27-4bd2-8d93-ec70804fbfed" (UID: "08c89ff9-7f27-4bd2-8d93-ec70804fbfed"). InnerVolumeSpecName "kube-api-access-v2b6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.273562 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b35f118-5646-441c-a833-ba6fb427dfdb-kube-api-access-p82hl" (OuterVolumeSpecName: "kube-api-access-p82hl") pod "5b35f118-5646-441c-a833-ba6fb427dfdb" (UID: "5b35f118-5646-441c-a833-ba6fb427dfdb"). InnerVolumeSpecName "kube-api-access-p82hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.280676 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b35f118-5646-441c-a833-ba6fb427dfdb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5b35f118-5646-441c-a833-ba6fb427dfdb" (UID: "5b35f118-5646-441c-a833-ba6fb427dfdb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.371496 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf109684-739e-4bdd-94bb-8ba8d975b187-serving-cert\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.371560 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-client-ca\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.371589 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-config\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.371617 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8djtk\" (UniqueName: \"kubernetes.io/projected/bf109684-739e-4bdd-94bb-8ba8d975b187-kube-api-access-8djtk\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.371641 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-proxy-ca-bundles\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.371734 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p82hl\" (UniqueName: \"kubernetes.io/projected/5b35f118-5646-441c-a833-ba6fb427dfdb-kube-api-access-p82hl\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.371750 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b35f118-5646-441c-a833-ba6fb427dfdb-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.371762 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.371773 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2b6j\" (UniqueName: \"kubernetes.io/projected/08c89ff9-7f27-4bd2-8d93-ec70804fbfed-kube-api-access-v2b6j\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.373247 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-client-ca\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.373934 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-config\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.374512 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-proxy-ca-bundles\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.377483 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf109684-739e-4bdd-94bb-8ba8d975b187-serving-cert\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.388571 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8djtk\" (UniqueName: \"kubernetes.io/projected/bf109684-739e-4bdd-94bb-8ba8d975b187-kube-api-access-8djtk\") pod \"controller-manager-7bcdf8b5f6-6mgj4\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.492012 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.800804 4805 generic.go:334] "Generic (PLEG): container finished" podID="08c89ff9-7f27-4bd2-8d93-ec70804fbfed" containerID="e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d" exitCode=0 Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.800871 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" event={"ID":"08c89ff9-7f27-4bd2-8d93-ec70804fbfed","Type":"ContainerDied","Data":"e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d"} Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.800879 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.800907 4805 scope.go:117] "RemoveContainer" containerID="e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.800897 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr" event={"ID":"08c89ff9-7f27-4bd2-8d93-ec70804fbfed","Type":"ContainerDied","Data":"8b4b083a5be887f825ec6df59d2cb5e6244fbffd9e951cd961fda009c10c92d9"} Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.802575 4805 generic.go:334] "Generic (PLEG): container finished" podID="5b35f118-5646-441c-a833-ba6fb427dfdb" containerID="db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e" exitCode=0 Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.802615 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" event={"ID":"5b35f118-5646-441c-a833-ba6fb427dfdb","Type":"ContainerDied","Data":"db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e"} Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.802632 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" event={"ID":"5b35f118-5646-441c-a833-ba6fb427dfdb","Type":"ContainerDied","Data":"6efc8768774bef1653ab860aa16067a73940ef8d4e98ae60462e9bda84195088"} Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.802673 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79794f6f4c-q65v6" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.824851 4805 scope.go:117] "RemoveContainer" containerID="e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d" Nov 28 15:32:09 crc kubenswrapper[4805]: E1128 15:32:09.827895 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d\": container with ID starting with e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d not found: ID does not exist" containerID="e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.827957 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d"} err="failed to get container status \"e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d\": rpc error: code = NotFound desc = could not find container \"e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d\": container with ID starting with e287e1fda2f75ebd82a1e8394efa6a74997318f2613ffb41d446f0338502304d not found: ID does not exist" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.827983 4805 scope.go:117] "RemoveContainer" containerID="db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.844822 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-79794f6f4c-q65v6"] Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.846928 4805 scope.go:117] "RemoveContainer" containerID="db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e" Nov 28 15:32:09 crc kubenswrapper[4805]: E1128 15:32:09.848701 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e\": container with ID starting with db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e not found: ID does not exist" containerID="db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.848747 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e"} err="failed to get container status \"db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e\": rpc error: code = NotFound desc = could not find container \"db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e\": container with ID starting with db85e11350c51d5f3f204b91d8050213520847c0742cd1248666b005dcef349e not found: ID does not exist" Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.850131 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-79794f6f4c-q65v6"] Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.853120 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr"] Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.857184 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c85c56bf-pzjjr"] Nov 28 15:32:09 crc kubenswrapper[4805]: I1128 15:32:09.882381 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4"] Nov 28 15:32:09 crc kubenswrapper[4805]: W1128 15:32:09.888070 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf109684_739e_4bdd_94bb_8ba8d975b187.slice/crio-4ad706493ff88a20da5d16e597de441b137350d2424fe84ae64676518b99467c WatchSource:0}: Error finding container 4ad706493ff88a20da5d16e597de441b137350d2424fe84ae64676518b99467c: Status 404 returned error can't find the container with id 4ad706493ff88a20da5d16e597de441b137350d2424fe84ae64676518b99467c Nov 28 15:32:10 crc kubenswrapper[4805]: I1128 15:32:10.809681 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" event={"ID":"bf109684-739e-4bdd-94bb-8ba8d975b187","Type":"ContainerStarted","Data":"e4fc7cc52b34eef43f239be9485bcca148808a853f43ab1b93ad8c909017440d"} Nov 28 15:32:10 crc kubenswrapper[4805]: I1128 15:32:10.809734 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" event={"ID":"bf109684-739e-4bdd-94bb-8ba8d975b187","Type":"ContainerStarted","Data":"4ad706493ff88a20da5d16e597de441b137350d2424fe84ae64676518b99467c"} Nov 28 15:32:10 crc kubenswrapper[4805]: I1128 15:32:10.809907 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:10 crc kubenswrapper[4805]: I1128 15:32:10.815339 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:10 crc kubenswrapper[4805]: I1128 15:32:10.823183 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" podStartSLOduration=3.8231638549999998 podStartE2EDuration="3.823163855s" podCreationTimestamp="2025-11-28 15:32:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:32:10.82264351 +0000 UTC m=+357.872434831" watchObservedRunningTime="2025-11-28 15:32:10.823163855 +0000 UTC m=+357.872955166" Nov 28 15:32:11 crc kubenswrapper[4805]: I1128 15:32:11.060541 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:32:11 crc kubenswrapper[4805]: I1128 15:32:11.060850 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:32:11 crc kubenswrapper[4805]: I1128 15:32:11.213057 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08c89ff9-7f27-4bd2-8d93-ec70804fbfed" path="/var/lib/kubelet/pods/08c89ff9-7f27-4bd2-8d93-ec70804fbfed/volumes" Nov 28 15:32:11 crc kubenswrapper[4805]: I1128 15:32:11.214020 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b35f118-5646-441c-a833-ba6fb427dfdb" path="/var/lib/kubelet/pods/5b35f118-5646-441c-a833-ba6fb427dfdb/volumes" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.027016 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl"] Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.028115 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.031106 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.031297 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.031374 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.031437 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.031975 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.033375 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.044975 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl"] Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.106842 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/571bb7a4-df08-48a1-a33d-7016ec74f9fb-serving-cert\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.106944 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-client-ca\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.106990 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-config\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.107050 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q6cn\" (UniqueName: \"kubernetes.io/projected/571bb7a4-df08-48a1-a33d-7016ec74f9fb-kube-api-access-8q6cn\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.207855 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/571bb7a4-df08-48a1-a33d-7016ec74f9fb-serving-cert\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.207910 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-client-ca\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.207939 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-config\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.207976 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q6cn\" (UniqueName: \"kubernetes.io/projected/571bb7a4-df08-48a1-a33d-7016ec74f9fb-kube-api-access-8q6cn\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.209342 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-client-ca\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.209462 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-config\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.215519 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/571bb7a4-df08-48a1-a33d-7016ec74f9fb-serving-cert\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.226399 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q6cn\" (UniqueName: \"kubernetes.io/projected/571bb7a4-df08-48a1-a33d-7016ec74f9fb-kube-api-access-8q6cn\") pod \"route-controller-manager-685f6d45b6-6lqwl\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.408830 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.600287 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl"] Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.823240 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" event={"ID":"571bb7a4-df08-48a1-a33d-7016ec74f9fb","Type":"ContainerStarted","Data":"2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8"} Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.823285 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" event={"ID":"571bb7a4-df08-48a1-a33d-7016ec74f9fb","Type":"ContainerStarted","Data":"5b9700c16dd4b1d7fef9fa697f4e8ab412eab4d78a17a319af4e43328c55d439"} Nov 28 15:32:12 crc kubenswrapper[4805]: I1128 15:32:12.840431 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" podStartSLOduration=5.840417426 podStartE2EDuration="5.840417426s" podCreationTimestamp="2025-11-28 15:32:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:32:12.836501024 +0000 UTC m=+359.886292335" watchObservedRunningTime="2025-11-28 15:32:12.840417426 +0000 UTC m=+359.890208737" Nov 28 15:32:13 crc kubenswrapper[4805]: I1128 15:32:13.828031 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:13 crc kubenswrapper[4805]: I1128 15:32:13.833015 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:16 crc kubenswrapper[4805]: I1128 15:32:16.917495 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jgwwr"] Nov 28 15:32:25 crc kubenswrapper[4805]: I1128 15:32:25.839729 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl"] Nov 28 15:32:25 crc kubenswrapper[4805]: I1128 15:32:25.840971 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" podUID="571bb7a4-df08-48a1-a33d-7016ec74f9fb" containerName="route-controller-manager" containerID="cri-o://2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8" gracePeriod=30 Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.848419 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.890647 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn"] Nov 28 15:32:26 crc kubenswrapper[4805]: E1128 15:32:26.890899 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="571bb7a4-df08-48a1-a33d-7016ec74f9fb" containerName="route-controller-manager" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.890918 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="571bb7a4-df08-48a1-a33d-7016ec74f9fb" containerName="route-controller-manager" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.891032 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="571bb7a4-df08-48a1-a33d-7016ec74f9fb" containerName="route-controller-manager" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.891504 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.901749 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn"] Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.905410 4805 generic.go:334] "Generic (PLEG): container finished" podID="571bb7a4-df08-48a1-a33d-7016ec74f9fb" containerID="2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8" exitCode=0 Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.905454 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" event={"ID":"571bb7a4-df08-48a1-a33d-7016ec74f9fb","Type":"ContainerDied","Data":"2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8"} Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.905481 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" event={"ID":"571bb7a4-df08-48a1-a33d-7016ec74f9fb","Type":"ContainerDied","Data":"5b9700c16dd4b1d7fef9fa697f4e8ab412eab4d78a17a319af4e43328c55d439"} Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.905498 4805 scope.go:117] "RemoveContainer" containerID="2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.905620 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.907218 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac079240-1b08-4189-bec4-6a93e1c9104f-config\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.907244 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac079240-1b08-4189-bec4-6a93e1c9104f-client-ca\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.907272 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac079240-1b08-4189-bec4-6a93e1c9104f-serving-cert\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.907308 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf49w\" (UniqueName: \"kubernetes.io/projected/ac079240-1b08-4189-bec4-6a93e1c9104f-kube-api-access-kf49w\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.920632 4805 scope.go:117] "RemoveContainer" containerID="2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8" Nov 28 15:32:26 crc kubenswrapper[4805]: E1128 15:32:26.921115 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8\": container with ID starting with 2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8 not found: ID does not exist" containerID="2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8" Nov 28 15:32:26 crc kubenswrapper[4805]: I1128 15:32:26.921151 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8"} err="failed to get container status \"2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8\": rpc error: code = NotFound desc = could not find container \"2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8\": container with ID starting with 2f76b155a7e3029fa0b848c77798c89a359fb99313096ed0ce5cbfbf9706d2f8 not found: ID does not exist" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.008075 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-config\") pod \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.008425 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/571bb7a4-df08-48a1-a33d-7016ec74f9fb-serving-cert\") pod \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.008489 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-client-ca\") pod \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.008510 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q6cn\" (UniqueName: \"kubernetes.io/projected/571bb7a4-df08-48a1-a33d-7016ec74f9fb-kube-api-access-8q6cn\") pod \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\" (UID: \"571bb7a4-df08-48a1-a33d-7016ec74f9fb\") " Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.008660 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac079240-1b08-4189-bec4-6a93e1c9104f-config\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.008693 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac079240-1b08-4189-bec4-6a93e1c9104f-client-ca\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.008726 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac079240-1b08-4189-bec4-6a93e1c9104f-serving-cert\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.008771 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf49w\" (UniqueName: \"kubernetes.io/projected/ac079240-1b08-4189-bec4-6a93e1c9104f-kube-api-access-kf49w\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.009166 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-client-ca" (OuterVolumeSpecName: "client-ca") pod "571bb7a4-df08-48a1-a33d-7016ec74f9fb" (UID: "571bb7a4-df08-48a1-a33d-7016ec74f9fb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.009182 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-config" (OuterVolumeSpecName: "config") pod "571bb7a4-df08-48a1-a33d-7016ec74f9fb" (UID: "571bb7a4-df08-48a1-a33d-7016ec74f9fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.009921 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac079240-1b08-4189-bec4-6a93e1c9104f-client-ca\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.010668 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac079240-1b08-4189-bec4-6a93e1c9104f-config\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.014147 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/571bb7a4-df08-48a1-a33d-7016ec74f9fb-kube-api-access-8q6cn" (OuterVolumeSpecName: "kube-api-access-8q6cn") pod "571bb7a4-df08-48a1-a33d-7016ec74f9fb" (UID: "571bb7a4-df08-48a1-a33d-7016ec74f9fb"). InnerVolumeSpecName "kube-api-access-8q6cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.014799 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac079240-1b08-4189-bec4-6a93e1c9104f-serving-cert\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.015338 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/571bb7a4-df08-48a1-a33d-7016ec74f9fb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "571bb7a4-df08-48a1-a33d-7016ec74f9fb" (UID: "571bb7a4-df08-48a1-a33d-7016ec74f9fb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.043027 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf49w\" (UniqueName: \"kubernetes.io/projected/ac079240-1b08-4189-bec4-6a93e1c9104f-kube-api-access-kf49w\") pod \"route-controller-manager-58c85c56bf-vctnn\" (UID: \"ac079240-1b08-4189-bec4-6a93e1c9104f\") " pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.109473 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.110295 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q6cn\" (UniqueName: \"kubernetes.io/projected/571bb7a4-df08-48a1-a33d-7016ec74f9fb-kube-api-access-8q6cn\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.110323 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/571bb7a4-df08-48a1-a33d-7016ec74f9fb-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.110336 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/571bb7a4-df08-48a1-a33d-7016ec74f9fb-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.213336 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.289939 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl"] Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.294550 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-6lqwl"] Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.684406 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn"] Nov 28 15:32:27 crc kubenswrapper[4805]: W1128 15:32:27.694587 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac079240_1b08_4189_bec4_6a93e1c9104f.slice/crio-9f283a426836292b6344dbfaae2cabf4eb4fb6bbfc41a508b37f3256282ac9b1 WatchSource:0}: Error finding container 9f283a426836292b6344dbfaae2cabf4eb4fb6bbfc41a508b37f3256282ac9b1: Status 404 returned error can't find the container with id 9f283a426836292b6344dbfaae2cabf4eb4fb6bbfc41a508b37f3256282ac9b1 Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.912934 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" event={"ID":"ac079240-1b08-4189-bec4-6a93e1c9104f","Type":"ContainerStarted","Data":"993f300e275ddcea4fdd016fc287bbda1e9337ec4a45b9ce1dab2c6c51510315"} Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.913383 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.913401 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" event={"ID":"ac079240-1b08-4189-bec4-6a93e1c9104f","Type":"ContainerStarted","Data":"9f283a426836292b6344dbfaae2cabf4eb4fb6bbfc41a508b37f3256282ac9b1"} Nov 28 15:32:27 crc kubenswrapper[4805]: I1128 15:32:27.930105 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" podStartSLOduration=2.930082468 podStartE2EDuration="2.930082468s" podCreationTimestamp="2025-11-28 15:32:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:32:27.929286285 +0000 UTC m=+374.979077596" watchObservedRunningTime="2025-11-28 15:32:27.930082468 +0000 UTC m=+374.979873799" Nov 28 15:32:28 crc kubenswrapper[4805]: I1128 15:32:28.338191 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-58c85c56bf-vctnn" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.141388 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bznbp"] Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.142253 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.160421 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bznbp"] Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.209226 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="571bb7a4-df08-48a1-a33d-7016ec74f9fb" path="/var/lib/kubelet/pods/571bb7a4-df08-48a1-a33d-7016ec74f9fb/volumes" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.244451 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/237e1b30-086a-4aa2-9971-ae18e728e602-registry-tls\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.244514 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/237e1b30-086a-4aa2-9971-ae18e728e602-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.244597 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/237e1b30-086a-4aa2-9971-ae18e728e602-trusted-ca\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.244639 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/237e1b30-086a-4aa2-9971-ae18e728e602-registry-certificates\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.244666 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/237e1b30-086a-4aa2-9971-ae18e728e602-bound-sa-token\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.244708 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/237e1b30-086a-4aa2-9971-ae18e728e602-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.244780 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.244810 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t2xh\" (UniqueName: \"kubernetes.io/projected/237e1b30-086a-4aa2-9971-ae18e728e602-kube-api-access-8t2xh\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.268147 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.346203 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/237e1b30-086a-4aa2-9971-ae18e728e602-trusted-ca\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.346264 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/237e1b30-086a-4aa2-9971-ae18e728e602-registry-certificates\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.346282 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/237e1b30-086a-4aa2-9971-ae18e728e602-bound-sa-token\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.346314 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/237e1b30-086a-4aa2-9971-ae18e728e602-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.346346 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t2xh\" (UniqueName: \"kubernetes.io/projected/237e1b30-086a-4aa2-9971-ae18e728e602-kube-api-access-8t2xh\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.346395 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/237e1b30-086a-4aa2-9971-ae18e728e602-registry-tls\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.346426 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/237e1b30-086a-4aa2-9971-ae18e728e602-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.346984 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/237e1b30-086a-4aa2-9971-ae18e728e602-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.348201 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/237e1b30-086a-4aa2-9971-ae18e728e602-registry-certificates\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.348201 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/237e1b30-086a-4aa2-9971-ae18e728e602-trusted-ca\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.351813 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/237e1b30-086a-4aa2-9971-ae18e728e602-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.353982 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/237e1b30-086a-4aa2-9971-ae18e728e602-registry-tls\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.364908 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t2xh\" (UniqueName: \"kubernetes.io/projected/237e1b30-086a-4aa2-9971-ae18e728e602-kube-api-access-8t2xh\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.366752 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/237e1b30-086a-4aa2-9971-ae18e728e602-bound-sa-token\") pod \"image-registry-66df7c8f76-bznbp\" (UID: \"237e1b30-086a-4aa2-9971-ae18e728e602\") " pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.457677 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.885326 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bznbp"] Nov 28 15:32:29 crc kubenswrapper[4805]: W1128 15:32:29.894330 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod237e1b30_086a_4aa2_9971_ae18e728e602.slice/crio-1119647f8a865d44af39411607617a4362a62398869d9921f5e8e4dc101d4ecd WatchSource:0}: Error finding container 1119647f8a865d44af39411607617a4362a62398869d9921f5e8e4dc101d4ecd: Status 404 returned error can't find the container with id 1119647f8a865d44af39411607617a4362a62398869d9921f5e8e4dc101d4ecd Nov 28 15:32:29 crc kubenswrapper[4805]: I1128 15:32:29.925680 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" event={"ID":"237e1b30-086a-4aa2-9971-ae18e728e602","Type":"ContainerStarted","Data":"1119647f8a865d44af39411607617a4362a62398869d9921f5e8e4dc101d4ecd"} Nov 28 15:32:30 crc kubenswrapper[4805]: I1128 15:32:30.932142 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" event={"ID":"237e1b30-086a-4aa2-9971-ae18e728e602","Type":"ContainerStarted","Data":"9f36a4d3fa525f67c76dba2b31d45ff2d9c3c3cfd785c0795c6144809cb2f8dd"} Nov 28 15:32:30 crc kubenswrapper[4805]: I1128 15:32:30.932445 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:30 crc kubenswrapper[4805]: I1128 15:32:30.951530 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" podStartSLOduration=1.951514016 podStartE2EDuration="1.951514016s" podCreationTimestamp="2025-11-28 15:32:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:32:30.948261513 +0000 UTC m=+377.998052814" watchObservedRunningTime="2025-11-28 15:32:30.951514016 +0000 UTC m=+378.001305327" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.712718 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jpf2f"] Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.714088 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.717880 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.724584 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jpf2f"] Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.820897 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98700b4-acf9-4b6c-926f-1a3591a3118c-utilities\") pod \"community-operators-jpf2f\" (UID: \"f98700b4-acf9-4b6c-926f-1a3591a3118c\") " pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.820974 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbzfv\" (UniqueName: \"kubernetes.io/projected/f98700b4-acf9-4b6c-926f-1a3591a3118c-kube-api-access-gbzfv\") pod \"community-operators-jpf2f\" (UID: \"f98700b4-acf9-4b6c-926f-1a3591a3118c\") " pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.821072 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98700b4-acf9-4b6c-926f-1a3591a3118c-catalog-content\") pod \"community-operators-jpf2f\" (UID: \"f98700b4-acf9-4b6c-926f-1a3591a3118c\") " pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.917117 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4d7mg"] Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.918714 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.920466 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.922521 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98700b4-acf9-4b6c-926f-1a3591a3118c-catalog-content\") pod \"community-operators-jpf2f\" (UID: \"f98700b4-acf9-4b6c-926f-1a3591a3118c\") " pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.922582 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98700b4-acf9-4b6c-926f-1a3591a3118c-utilities\") pod \"community-operators-jpf2f\" (UID: \"f98700b4-acf9-4b6c-926f-1a3591a3118c\") " pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.922633 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbzfv\" (UniqueName: \"kubernetes.io/projected/f98700b4-acf9-4b6c-926f-1a3591a3118c-kube-api-access-gbzfv\") pod \"community-operators-jpf2f\" (UID: \"f98700b4-acf9-4b6c-926f-1a3591a3118c\") " pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.923289 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98700b4-acf9-4b6c-926f-1a3591a3118c-catalog-content\") pod \"community-operators-jpf2f\" (UID: \"f98700b4-acf9-4b6c-926f-1a3591a3118c\") " pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.923324 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98700b4-acf9-4b6c-926f-1a3591a3118c-utilities\") pod \"community-operators-jpf2f\" (UID: \"f98700b4-acf9-4b6c-926f-1a3591a3118c\") " pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.939753 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4d7mg"] Nov 28 15:32:34 crc kubenswrapper[4805]: I1128 15:32:34.957271 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbzfv\" (UniqueName: \"kubernetes.io/projected/f98700b4-acf9-4b6c-926f-1a3591a3118c-kube-api-access-gbzfv\") pod \"community-operators-jpf2f\" (UID: \"f98700b4-acf9-4b6c-926f-1a3591a3118c\") " pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.023723 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e223a05-6a95-4786-a73e-3cfbc00188b9-catalog-content\") pod \"certified-operators-4d7mg\" (UID: \"1e223a05-6a95-4786-a73e-3cfbc00188b9\") " pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.024088 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gfww\" (UniqueName: \"kubernetes.io/projected/1e223a05-6a95-4786-a73e-3cfbc00188b9-kube-api-access-6gfww\") pod \"certified-operators-4d7mg\" (UID: \"1e223a05-6a95-4786-a73e-3cfbc00188b9\") " pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.024105 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e223a05-6a95-4786-a73e-3cfbc00188b9-utilities\") pod \"certified-operators-4d7mg\" (UID: \"1e223a05-6a95-4786-a73e-3cfbc00188b9\") " pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.031667 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.124860 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e223a05-6a95-4786-a73e-3cfbc00188b9-catalog-content\") pod \"certified-operators-4d7mg\" (UID: \"1e223a05-6a95-4786-a73e-3cfbc00188b9\") " pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.124934 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gfww\" (UniqueName: \"kubernetes.io/projected/1e223a05-6a95-4786-a73e-3cfbc00188b9-kube-api-access-6gfww\") pod \"certified-operators-4d7mg\" (UID: \"1e223a05-6a95-4786-a73e-3cfbc00188b9\") " pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.124954 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e223a05-6a95-4786-a73e-3cfbc00188b9-utilities\") pod \"certified-operators-4d7mg\" (UID: \"1e223a05-6a95-4786-a73e-3cfbc00188b9\") " pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.125343 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e223a05-6a95-4786-a73e-3cfbc00188b9-utilities\") pod \"certified-operators-4d7mg\" (UID: \"1e223a05-6a95-4786-a73e-3cfbc00188b9\") " pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.125420 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e223a05-6a95-4786-a73e-3cfbc00188b9-catalog-content\") pod \"certified-operators-4d7mg\" (UID: \"1e223a05-6a95-4786-a73e-3cfbc00188b9\") " pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.151315 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gfww\" (UniqueName: \"kubernetes.io/projected/1e223a05-6a95-4786-a73e-3cfbc00188b9-kube-api-access-6gfww\") pod \"certified-operators-4d7mg\" (UID: \"1e223a05-6a95-4786-a73e-3cfbc00188b9\") " pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.253222 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.476212 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jpf2f"] Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.659232 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4d7mg"] Nov 28 15:32:35 crc kubenswrapper[4805]: W1128 15:32:35.686237 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e223a05_6a95_4786_a73e_3cfbc00188b9.slice/crio-dce3be78a04d92d081187769a37c1d25109298eaf5c729cea1fb44771a428c33 WatchSource:0}: Error finding container dce3be78a04d92d081187769a37c1d25109298eaf5c729cea1fb44771a428c33: Status 404 returned error can't find the container with id dce3be78a04d92d081187769a37c1d25109298eaf5c729cea1fb44771a428c33 Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.962518 4805 generic.go:334] "Generic (PLEG): container finished" podID="1e223a05-6a95-4786-a73e-3cfbc00188b9" containerID="39dfd345bcd0a463aac5e5ba46b8ef8f5c59d6aa5ed9374762db2de75e9c2424" exitCode=0 Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.962592 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d7mg" event={"ID":"1e223a05-6a95-4786-a73e-3cfbc00188b9","Type":"ContainerDied","Data":"39dfd345bcd0a463aac5e5ba46b8ef8f5c59d6aa5ed9374762db2de75e9c2424"} Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.962928 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d7mg" event={"ID":"1e223a05-6a95-4786-a73e-3cfbc00188b9","Type":"ContainerStarted","Data":"dce3be78a04d92d081187769a37c1d25109298eaf5c729cea1fb44771a428c33"} Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.964889 4805 generic.go:334] "Generic (PLEG): container finished" podID="f98700b4-acf9-4b6c-926f-1a3591a3118c" containerID="bc647e3d4ff882479de1360d64f8fb124cbfe8807e4fef1856914c77d8ef20f1" exitCode=0 Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.964917 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpf2f" event={"ID":"f98700b4-acf9-4b6c-926f-1a3591a3118c","Type":"ContainerDied","Data":"bc647e3d4ff882479de1360d64f8fb124cbfe8807e4fef1856914c77d8ef20f1"} Nov 28 15:32:35 crc kubenswrapper[4805]: I1128 15:32:35.964939 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpf2f" event={"ID":"f98700b4-acf9-4b6c-926f-1a3591a3118c","Type":"ContainerStarted","Data":"029537cf1c7c865ab9732aa7dc94a46d21119448685608ef0366dfdc2035f673"} Nov 28 15:32:36 crc kubenswrapper[4805]: I1128 15:32:36.974005 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d7mg" event={"ID":"1e223a05-6a95-4786-a73e-3cfbc00188b9","Type":"ContainerStarted","Data":"b77d138ea83c041bfeda1431f8d132213999c9596658f9bbbdff72be4247ca1a"} Nov 28 15:32:36 crc kubenswrapper[4805]: I1128 15:32:36.976649 4805 generic.go:334] "Generic (PLEG): container finished" podID="f98700b4-acf9-4b6c-926f-1a3591a3118c" containerID="31e46f2f62babf4bb790f16e497815590f263cd870dfd8d68e70e9f3ff0ebe41" exitCode=0 Nov 28 15:32:36 crc kubenswrapper[4805]: I1128 15:32:36.976697 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpf2f" event={"ID":"f98700b4-acf9-4b6c-926f-1a3591a3118c","Type":"ContainerDied","Data":"31e46f2f62babf4bb790f16e497815590f263cd870dfd8d68e70e9f3ff0ebe41"} Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.106697 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c4tws"] Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.107915 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.109390 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.113929 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4tws"] Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.249409 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29adeed6-df07-497c-9d0f-6a8b818448c8-catalog-content\") pod \"redhat-marketplace-c4tws\" (UID: \"29adeed6-df07-497c-9d0f-6a8b818448c8\") " pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.249541 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29adeed6-df07-497c-9d0f-6a8b818448c8-utilities\") pod \"redhat-marketplace-c4tws\" (UID: \"29adeed6-df07-497c-9d0f-6a8b818448c8\") " pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.249582 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpt5m\" (UniqueName: \"kubernetes.io/projected/29adeed6-df07-497c-9d0f-6a8b818448c8-kube-api-access-zpt5m\") pod \"redhat-marketplace-c4tws\" (UID: \"29adeed6-df07-497c-9d0f-6a8b818448c8\") " pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.311104 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n77t9"] Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.313631 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.322140 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.323785 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n77t9"] Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.351160 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpt5m\" (UniqueName: \"kubernetes.io/projected/29adeed6-df07-497c-9d0f-6a8b818448c8-kube-api-access-zpt5m\") pod \"redhat-marketplace-c4tws\" (UID: \"29adeed6-df07-497c-9d0f-6a8b818448c8\") " pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.351200 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29adeed6-df07-497c-9d0f-6a8b818448c8-catalog-content\") pod \"redhat-marketplace-c4tws\" (UID: \"29adeed6-df07-497c-9d0f-6a8b818448c8\") " pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.351270 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29adeed6-df07-497c-9d0f-6a8b818448c8-utilities\") pod \"redhat-marketplace-c4tws\" (UID: \"29adeed6-df07-497c-9d0f-6a8b818448c8\") " pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.352107 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29adeed6-df07-497c-9d0f-6a8b818448c8-utilities\") pod \"redhat-marketplace-c4tws\" (UID: \"29adeed6-df07-497c-9d0f-6a8b818448c8\") " pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.352455 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29adeed6-df07-497c-9d0f-6a8b818448c8-catalog-content\") pod \"redhat-marketplace-c4tws\" (UID: \"29adeed6-df07-497c-9d0f-6a8b818448c8\") " pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.376172 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpt5m\" (UniqueName: \"kubernetes.io/projected/29adeed6-df07-497c-9d0f-6a8b818448c8-kube-api-access-zpt5m\") pod \"redhat-marketplace-c4tws\" (UID: \"29adeed6-df07-497c-9d0f-6a8b818448c8\") " pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.424301 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.451969 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zd5m\" (UniqueName: \"kubernetes.io/projected/17c1bdf1-aa1e-4857-9708-0c0382cd70fd-kube-api-access-6zd5m\") pod \"redhat-operators-n77t9\" (UID: \"17c1bdf1-aa1e-4857-9708-0c0382cd70fd\") " pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.452082 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17c1bdf1-aa1e-4857-9708-0c0382cd70fd-utilities\") pod \"redhat-operators-n77t9\" (UID: \"17c1bdf1-aa1e-4857-9708-0c0382cd70fd\") " pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.452135 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17c1bdf1-aa1e-4857-9708-0c0382cd70fd-catalog-content\") pod \"redhat-operators-n77t9\" (UID: \"17c1bdf1-aa1e-4857-9708-0c0382cd70fd\") " pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.553946 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zd5m\" (UniqueName: \"kubernetes.io/projected/17c1bdf1-aa1e-4857-9708-0c0382cd70fd-kube-api-access-6zd5m\") pod \"redhat-operators-n77t9\" (UID: \"17c1bdf1-aa1e-4857-9708-0c0382cd70fd\") " pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.554313 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17c1bdf1-aa1e-4857-9708-0c0382cd70fd-utilities\") pod \"redhat-operators-n77t9\" (UID: \"17c1bdf1-aa1e-4857-9708-0c0382cd70fd\") " pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.554379 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17c1bdf1-aa1e-4857-9708-0c0382cd70fd-catalog-content\") pod \"redhat-operators-n77t9\" (UID: \"17c1bdf1-aa1e-4857-9708-0c0382cd70fd\") " pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.555244 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17c1bdf1-aa1e-4857-9708-0c0382cd70fd-catalog-content\") pod \"redhat-operators-n77t9\" (UID: \"17c1bdf1-aa1e-4857-9708-0c0382cd70fd\") " pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.555651 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17c1bdf1-aa1e-4857-9708-0c0382cd70fd-utilities\") pod \"redhat-operators-n77t9\" (UID: \"17c1bdf1-aa1e-4857-9708-0c0382cd70fd\") " pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.576941 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zd5m\" (UniqueName: \"kubernetes.io/projected/17c1bdf1-aa1e-4857-9708-0c0382cd70fd-kube-api-access-6zd5m\") pod \"redhat-operators-n77t9\" (UID: \"17c1bdf1-aa1e-4857-9708-0c0382cd70fd\") " pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.637480 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.867984 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4tws"] Nov 28 15:32:37 crc kubenswrapper[4805]: W1128 15:32:37.876888 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29adeed6_df07_497c_9d0f_6a8b818448c8.slice/crio-cd91078ef535d8488e07874cb7f3920a4306d0b32fe475668acf4f6056adb659 WatchSource:0}: Error finding container cd91078ef535d8488e07874cb7f3920a4306d0b32fe475668acf4f6056adb659: Status 404 returned error can't find the container with id cd91078ef535d8488e07874cb7f3920a4306d0b32fe475668acf4f6056adb659 Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.983209 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4tws" event={"ID":"29adeed6-df07-497c-9d0f-6a8b818448c8","Type":"ContainerStarted","Data":"cd91078ef535d8488e07874cb7f3920a4306d0b32fe475668acf4f6056adb659"} Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.985503 4805 generic.go:334] "Generic (PLEG): container finished" podID="1e223a05-6a95-4786-a73e-3cfbc00188b9" containerID="b77d138ea83c041bfeda1431f8d132213999c9596658f9bbbdff72be4247ca1a" exitCode=0 Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.985563 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d7mg" event={"ID":"1e223a05-6a95-4786-a73e-3cfbc00188b9","Type":"ContainerDied","Data":"b77d138ea83c041bfeda1431f8d132213999c9596658f9bbbdff72be4247ca1a"} Nov 28 15:32:37 crc kubenswrapper[4805]: I1128 15:32:37.988126 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpf2f" event={"ID":"f98700b4-acf9-4b6c-926f-1a3591a3118c","Type":"ContainerStarted","Data":"c9cf9aac9db91dfb9463404725809ab807194467bd06b32bd1a5a855888de066"} Nov 28 15:32:38 crc kubenswrapper[4805]: I1128 15:32:38.023644 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jpf2f" podStartSLOduration=2.605991365 podStartE2EDuration="4.023629423s" podCreationTimestamp="2025-11-28 15:32:34 +0000 UTC" firstStartedPulling="2025-11-28 15:32:35.966018762 +0000 UTC m=+383.015810113" lastFinishedPulling="2025-11-28 15:32:37.38365686 +0000 UTC m=+384.433448171" observedRunningTime="2025-11-28 15:32:38.021768269 +0000 UTC m=+385.071559580" watchObservedRunningTime="2025-11-28 15:32:38.023629423 +0000 UTC m=+385.073420734" Nov 28 15:32:38 crc kubenswrapper[4805]: I1128 15:32:38.043211 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n77t9"] Nov 28 15:32:38 crc kubenswrapper[4805]: I1128 15:32:38.995721 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d7mg" event={"ID":"1e223a05-6a95-4786-a73e-3cfbc00188b9","Type":"ContainerStarted","Data":"5a01aab148eced6ab451294ce75b56fa992bffe738bc46be8d25e4d7cdb00d35"} Nov 28 15:32:38 crc kubenswrapper[4805]: I1128 15:32:38.998739 4805 generic.go:334] "Generic (PLEG): container finished" podID="17c1bdf1-aa1e-4857-9708-0c0382cd70fd" containerID="7da9a0a3eb4bacf1ba05614d7679a407d836448ee62a400fa8a2e738050e3726" exitCode=0 Nov 28 15:32:38 crc kubenswrapper[4805]: I1128 15:32:38.998820 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n77t9" event={"ID":"17c1bdf1-aa1e-4857-9708-0c0382cd70fd","Type":"ContainerDied","Data":"7da9a0a3eb4bacf1ba05614d7679a407d836448ee62a400fa8a2e738050e3726"} Nov 28 15:32:38 crc kubenswrapper[4805]: I1128 15:32:38.998855 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n77t9" event={"ID":"17c1bdf1-aa1e-4857-9708-0c0382cd70fd","Type":"ContainerStarted","Data":"6c298790cd4ce64d284c3c6985d5bacc417181d5a6f3f3c2b9cd2bdee904a767"} Nov 28 15:32:39 crc kubenswrapper[4805]: I1128 15:32:39.000169 4805 generic.go:334] "Generic (PLEG): container finished" podID="29adeed6-df07-497c-9d0f-6a8b818448c8" containerID="5c04f66dee5ba50e09a4a60432c80e261ba07288302c390cc64ec906505de0aa" exitCode=0 Nov 28 15:32:39 crc kubenswrapper[4805]: I1128 15:32:39.000205 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4tws" event={"ID":"29adeed6-df07-497c-9d0f-6a8b818448c8","Type":"ContainerDied","Data":"5c04f66dee5ba50e09a4a60432c80e261ba07288302c390cc64ec906505de0aa"} Nov 28 15:32:39 crc kubenswrapper[4805]: I1128 15:32:39.014341 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4d7mg" podStartSLOduration=2.575292154 podStartE2EDuration="5.014322114s" podCreationTimestamp="2025-11-28 15:32:34 +0000 UTC" firstStartedPulling="2025-11-28 15:32:35.964928621 +0000 UTC m=+383.014719972" lastFinishedPulling="2025-11-28 15:32:38.403958591 +0000 UTC m=+385.453749932" observedRunningTime="2025-11-28 15:32:39.012926493 +0000 UTC m=+386.062717804" watchObservedRunningTime="2025-11-28 15:32:39.014322114 +0000 UTC m=+386.064113425" Nov 28 15:32:40 crc kubenswrapper[4805]: I1128 15:32:40.006221 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n77t9" event={"ID":"17c1bdf1-aa1e-4857-9708-0c0382cd70fd","Type":"ContainerStarted","Data":"f87fefa678a904bf14eadfbf232648aec61dfa896365b488fa5c2d6103e44894"} Nov 28 15:32:40 crc kubenswrapper[4805]: I1128 15:32:40.007902 4805 generic.go:334] "Generic (PLEG): container finished" podID="29adeed6-df07-497c-9d0f-6a8b818448c8" containerID="29850b13494645ca136591d76d0ea34d3f6c22ed216cefbb7dc60b3784dcd6a5" exitCode=0 Nov 28 15:32:40 crc kubenswrapper[4805]: I1128 15:32:40.008858 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4tws" event={"ID":"29adeed6-df07-497c-9d0f-6a8b818448c8","Type":"ContainerDied","Data":"29850b13494645ca136591d76d0ea34d3f6c22ed216cefbb7dc60b3784dcd6a5"} Nov 28 15:32:41 crc kubenswrapper[4805]: I1128 15:32:41.013635 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4tws" event={"ID":"29adeed6-df07-497c-9d0f-6a8b818448c8","Type":"ContainerStarted","Data":"8de3c09618e5d423afd4848e3b5131aa35b4307f514491077316ecb29cd549cf"} Nov 28 15:32:41 crc kubenswrapper[4805]: I1128 15:32:41.015965 4805 generic.go:334] "Generic (PLEG): container finished" podID="17c1bdf1-aa1e-4857-9708-0c0382cd70fd" containerID="f87fefa678a904bf14eadfbf232648aec61dfa896365b488fa5c2d6103e44894" exitCode=0 Nov 28 15:32:41 crc kubenswrapper[4805]: I1128 15:32:41.016000 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n77t9" event={"ID":"17c1bdf1-aa1e-4857-9708-0c0382cd70fd","Type":"ContainerDied","Data":"f87fefa678a904bf14eadfbf232648aec61dfa896365b488fa5c2d6103e44894"} Nov 28 15:32:41 crc kubenswrapper[4805]: I1128 15:32:41.058350 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c4tws" podStartSLOduration=2.3283991410000002 podStartE2EDuration="4.058334283s" podCreationTimestamp="2025-11-28 15:32:37 +0000 UTC" firstStartedPulling="2025-11-28 15:32:39.00130723 +0000 UTC m=+386.051098551" lastFinishedPulling="2025-11-28 15:32:40.731242382 +0000 UTC m=+387.781033693" observedRunningTime="2025-11-28 15:32:41.03766328 +0000 UTC m=+388.087454621" watchObservedRunningTime="2025-11-28 15:32:41.058334283 +0000 UTC m=+388.108125594" Nov 28 15:32:41 crc kubenswrapper[4805]: I1128 15:32:41.059935 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:32:41 crc kubenswrapper[4805]: I1128 15:32:41.060004 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:32:41 crc kubenswrapper[4805]: I1128 15:32:41.961956 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" podUID="aebbc881-3380-4df9-b082-f6e0d16b8f8e" containerName="oauth-openshift" containerID="cri-o://7b4663954d21b58e002fc6e305f94fd1b625d8112c41e6b5bfb0d34e95f6a00d" gracePeriod=15 Nov 28 15:32:42 crc kubenswrapper[4805]: I1128 15:32:42.024174 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n77t9" event={"ID":"17c1bdf1-aa1e-4857-9708-0c0382cd70fd","Type":"ContainerStarted","Data":"9e5e71dd159b77aa136eed096df7cf5b5cdad252cd9d18cf49a2ac521d3f4196"} Nov 28 15:32:42 crc kubenswrapper[4805]: I1128 15:32:42.048560 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n77t9" podStartSLOduration=2.54300632 podStartE2EDuration="5.0485382s" podCreationTimestamp="2025-11-28 15:32:37 +0000 UTC" firstStartedPulling="2025-11-28 15:32:39.000088584 +0000 UTC m=+386.049879905" lastFinishedPulling="2025-11-28 15:32:41.505620474 +0000 UTC m=+388.555411785" observedRunningTime="2025-11-28 15:32:42.04645504 +0000 UTC m=+389.096246381" watchObservedRunningTime="2025-11-28 15:32:42.0485382 +0000 UTC m=+389.098329521" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.057457 4805 generic.go:334] "Generic (PLEG): container finished" podID="aebbc881-3380-4df9-b082-f6e0d16b8f8e" containerID="7b4663954d21b58e002fc6e305f94fd1b625d8112c41e6b5bfb0d34e95f6a00d" exitCode=0 Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.057542 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" event={"ID":"aebbc881-3380-4df9-b082-f6e0d16b8f8e","Type":"ContainerDied","Data":"7b4663954d21b58e002fc6e305f94fd1b625d8112c41e6b5bfb0d34e95f6a00d"} Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.190472 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.221977 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6cb8f88875-qbbfg"] Nov 28 15:32:44 crc kubenswrapper[4805]: E1128 15:32:44.222215 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aebbc881-3380-4df9-b082-f6e0d16b8f8e" containerName="oauth-openshift" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.222229 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="aebbc881-3380-4df9-b082-f6e0d16b8f8e" containerName="oauth-openshift" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.222347 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="aebbc881-3380-4df9-b082-f6e0d16b8f8e" containerName="oauth-openshift" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.222784 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.242201 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cb8f88875-qbbfg"] Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.361762 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-trusted-ca-bundle\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.361815 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-idp-0-file-data\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.361867 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-error\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.361900 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-dir\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.361917 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-policies\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.361964 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-ocp-branding-template\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.361995 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-router-certs\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362011 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-service-ca\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362046 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-provider-selection\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362074 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvrhs\" (UniqueName: \"kubernetes.io/projected/aebbc881-3380-4df9-b082-f6e0d16b8f8e-kube-api-access-kvrhs\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362095 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-login\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362119 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-cliconfig\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362155 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-session\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362181 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-serving-cert\") pod \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\" (UID: \"aebbc881-3380-4df9-b082-f6e0d16b8f8e\") " Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362424 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-template-error\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362460 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-audit-policies\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362491 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-template-login\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362520 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsbmd\" (UniqueName: \"kubernetes.io/projected/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-kube-api-access-xsbmd\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362545 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-audit-dir\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362567 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362590 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362617 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362641 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362664 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362687 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362682 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.362710 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.363229 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.363292 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.363428 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.363515 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.363760 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.364004 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-session\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.364215 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.364231 4805 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.364243 4805 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.364257 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.364270 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.367905 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.368179 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.368610 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.370816 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.372690 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.372938 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.373352 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.373458 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aebbc881-3380-4df9-b082-f6e0d16b8f8e-kube-api-access-kvrhs" (OuterVolumeSpecName: "kube-api-access-kvrhs") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "kube-api-access-kvrhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.375747 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "aebbc881-3380-4df9-b082-f6e0d16b8f8e" (UID: "aebbc881-3380-4df9-b082-f6e0d16b8f8e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465705 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-audit-policies\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465767 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-template-login\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465797 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsbmd\" (UniqueName: \"kubernetes.io/projected/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-kube-api-access-xsbmd\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465817 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-audit-dir\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465831 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465848 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465870 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465888 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465903 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465921 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465947 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465965 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.465984 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-session\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466052 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-template-error\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466114 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466129 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466142 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466156 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466168 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466181 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466194 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvrhs\" (UniqueName: \"kubernetes.io/projected/aebbc881-3380-4df9-b082-f6e0d16b8f8e-kube-api-access-kvrhs\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466209 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466222 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aebbc881-3380-4df9-b082-f6e0d16b8f8e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.466787 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-audit-policies\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.467912 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-audit-dir\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.468193 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.470845 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.471148 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.471216 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-session\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.471255 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.472015 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.472554 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.473043 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.481085 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-template-error\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.481305 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-user-template-login\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.483246 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.488429 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsbmd\" (UniqueName: \"kubernetes.io/projected/7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf-kube-api-access-xsbmd\") pod \"oauth-openshift-6cb8f88875-qbbfg\" (UID: \"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf\") " pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.544548 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:44 crc kubenswrapper[4805]: I1128 15:32:44.948709 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cb8f88875-qbbfg"] Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.047759 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.047836 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.077067 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" event={"ID":"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf","Type":"ContainerStarted","Data":"b789a0dfeb8a67112bdeeeb5345584f05ee71b6ba385f99309fb802e9f53c8b1"} Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.079010 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" event={"ID":"aebbc881-3380-4df9-b082-f6e0d16b8f8e","Type":"ContainerDied","Data":"e9392d43ac9ded6ef3b6a7c15d003fa17493e8a1600ecb3dcaeae3a93513a456"} Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.079052 4805 scope.go:117] "RemoveContainer" containerID="7b4663954d21b58e002fc6e305f94fd1b625d8112c41e6b5bfb0d34e95f6a00d" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.079079 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jgwwr" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.099240 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.111961 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jgwwr"] Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.116077 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jgwwr"] Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.147989 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jpf2f" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.213804 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aebbc881-3380-4df9-b082-f6e0d16b8f8e" path="/var/lib/kubelet/pods/aebbc881-3380-4df9-b082-f6e0d16b8f8e/volumes" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.253651 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.253715 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.289905 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.815569 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4"] Nov 28 15:32:45 crc kubenswrapper[4805]: I1128 15:32:45.815818 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" podUID="bf109684-739e-4bdd-94bb-8ba8d975b187" containerName="controller-manager" containerID="cri-o://e4fc7cc52b34eef43f239be9485bcca148808a853f43ab1b93ad8c909017440d" gracePeriod=30 Nov 28 15:32:46 crc kubenswrapper[4805]: I1128 15:32:46.085195 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" event={"ID":"7bd9ff0a-80b0-4b50-84e4-ddeb6670b5cf","Type":"ContainerStarted","Data":"41eb25e1956ec309f5fd8f8682f7a941d29ffe24e9f53235a2da9861178c54fb"} Nov 28 15:32:46 crc kubenswrapper[4805]: I1128 15:32:46.085567 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:46 crc kubenswrapper[4805]: I1128 15:32:46.095597 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" Nov 28 15:32:46 crc kubenswrapper[4805]: I1128 15:32:46.108118 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6cb8f88875-qbbfg" podStartSLOduration=30.108102071 podStartE2EDuration="30.108102071s" podCreationTimestamp="2025-11-28 15:32:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:32:46.106711974 +0000 UTC m=+393.156503285" watchObservedRunningTime="2025-11-28 15:32:46.108102071 +0000 UTC m=+393.157893382" Nov 28 15:32:46 crc kubenswrapper[4805]: I1128 15:32:46.133154 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4d7mg" Nov 28 15:32:47 crc kubenswrapper[4805]: I1128 15:32:47.424646 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:47 crc kubenswrapper[4805]: I1128 15:32:47.426290 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:47 crc kubenswrapper[4805]: I1128 15:32:47.466251 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:47 crc kubenswrapper[4805]: I1128 15:32:47.638156 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:47 crc kubenswrapper[4805]: I1128 15:32:47.638215 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:47 crc kubenswrapper[4805]: I1128 15:32:47.685244 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:48 crc kubenswrapper[4805]: I1128 15:32:48.142229 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c4tws" Nov 28 15:32:48 crc kubenswrapper[4805]: I1128 15:32:48.162285 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n77t9" Nov 28 15:32:49 crc kubenswrapper[4805]: I1128 15:32:49.110395 4805 generic.go:334] "Generic (PLEG): container finished" podID="bf109684-739e-4bdd-94bb-8ba8d975b187" containerID="e4fc7cc52b34eef43f239be9485bcca148808a853f43ab1b93ad8c909017440d" exitCode=0 Nov 28 15:32:49 crc kubenswrapper[4805]: I1128 15:32:49.110502 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" event={"ID":"bf109684-739e-4bdd-94bb-8ba8d975b187","Type":"ContainerDied","Data":"e4fc7cc52b34eef43f239be9485bcca148808a853f43ab1b93ad8c909017440d"} Nov 28 15:32:49 crc kubenswrapper[4805]: I1128 15:32:49.466310 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bznbp" Nov 28 15:32:49 crc kubenswrapper[4805]: I1128 15:32:49.493037 4805 patch_prober.go:28] interesting pod/controller-manager-7bcdf8b5f6-6mgj4 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" start-of-body= Nov 28 15:32:49 crc kubenswrapper[4805]: I1128 15:32:49.493124 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" podUID="bf109684-739e-4bdd-94bb-8ba8d975b187" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" Nov 28 15:32:49 crc kubenswrapper[4805]: I1128 15:32:49.538152 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5vvq"] Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.069900 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.097339 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-79794f6f4c-zf854"] Nov 28 15:32:50 crc kubenswrapper[4805]: E1128 15:32:50.097582 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf109684-739e-4bdd-94bb-8ba8d975b187" containerName="controller-manager" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.097809 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf109684-739e-4bdd-94bb-8ba8d975b187" containerName="controller-manager" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.097961 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf109684-739e-4bdd-94bb-8ba8d975b187" containerName="controller-manager" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.098393 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.103650 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79794f6f4c-zf854"] Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.122808 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.122809 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4" event={"ID":"bf109684-739e-4bdd-94bb-8ba8d975b187","Type":"ContainerDied","Data":"4ad706493ff88a20da5d16e597de441b137350d2424fe84ae64676518b99467c"} Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.122952 4805 scope.go:117] "RemoveContainer" containerID="e4fc7cc52b34eef43f239be9485bcca148808a853f43ab1b93ad8c909017440d" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.137657 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8djtk\" (UniqueName: \"kubernetes.io/projected/bf109684-739e-4bdd-94bb-8ba8d975b187-kube-api-access-8djtk\") pod \"bf109684-739e-4bdd-94bb-8ba8d975b187\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.137740 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-proxy-ca-bundles\") pod \"bf109684-739e-4bdd-94bb-8ba8d975b187\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.137786 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-client-ca\") pod \"bf109684-739e-4bdd-94bb-8ba8d975b187\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.137802 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf109684-739e-4bdd-94bb-8ba8d975b187-serving-cert\") pod \"bf109684-739e-4bdd-94bb-8ba8d975b187\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.137853 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-config\") pod \"bf109684-739e-4bdd-94bb-8ba8d975b187\" (UID: \"bf109684-739e-4bdd-94bb-8ba8d975b187\") " Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.137991 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d57a7c0e-c79a-43da-a27b-f4b7f991908e-config\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.138022 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d57a7c0e-c79a-43da-a27b-f4b7f991908e-client-ca\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.138044 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d57a7c0e-c79a-43da-a27b-f4b7f991908e-serving-cert\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.138059 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lst2g\" (UniqueName: \"kubernetes.io/projected/d57a7c0e-c79a-43da-a27b-f4b7f991908e-kube-api-access-lst2g\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.138278 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d57a7c0e-c79a-43da-a27b-f4b7f991908e-proxy-ca-bundles\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.138627 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "bf109684-739e-4bdd-94bb-8ba8d975b187" (UID: "bf109684-739e-4bdd-94bb-8ba8d975b187"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.138648 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-client-ca" (OuterVolumeSpecName: "client-ca") pod "bf109684-739e-4bdd-94bb-8ba8d975b187" (UID: "bf109684-739e-4bdd-94bb-8ba8d975b187"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.138987 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-config" (OuterVolumeSpecName: "config") pod "bf109684-739e-4bdd-94bb-8ba8d975b187" (UID: "bf109684-739e-4bdd-94bb-8ba8d975b187"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.142989 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf109684-739e-4bdd-94bb-8ba8d975b187-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bf109684-739e-4bdd-94bb-8ba8d975b187" (UID: "bf109684-739e-4bdd-94bb-8ba8d975b187"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.155183 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf109684-739e-4bdd-94bb-8ba8d975b187-kube-api-access-8djtk" (OuterVolumeSpecName: "kube-api-access-8djtk") pod "bf109684-739e-4bdd-94bb-8ba8d975b187" (UID: "bf109684-739e-4bdd-94bb-8ba8d975b187"). InnerVolumeSpecName "kube-api-access-8djtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.239799 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d57a7c0e-c79a-43da-a27b-f4b7f991908e-config\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.239843 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d57a7c0e-c79a-43da-a27b-f4b7f991908e-client-ca\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.239866 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d57a7c0e-c79a-43da-a27b-f4b7f991908e-serving-cert\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.239882 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lst2g\" (UniqueName: \"kubernetes.io/projected/d57a7c0e-c79a-43da-a27b-f4b7f991908e-kube-api-access-lst2g\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.240103 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d57a7c0e-c79a-43da-a27b-f4b7f991908e-proxy-ca-bundles\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.240259 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.240273 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.240284 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf109684-739e-4bdd-94bb-8ba8d975b187-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.240297 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf109684-739e-4bdd-94bb-8ba8d975b187-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.240310 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8djtk\" (UniqueName: \"kubernetes.io/projected/bf109684-739e-4bdd-94bb-8ba8d975b187-kube-api-access-8djtk\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.241109 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d57a7c0e-c79a-43da-a27b-f4b7f991908e-proxy-ca-bundles\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.241196 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d57a7c0e-c79a-43da-a27b-f4b7f991908e-client-ca\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.241284 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d57a7c0e-c79a-43da-a27b-f4b7f991908e-config\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.244244 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d57a7c0e-c79a-43da-a27b-f4b7f991908e-serving-cert\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.260477 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lst2g\" (UniqueName: \"kubernetes.io/projected/d57a7c0e-c79a-43da-a27b-f4b7f991908e-kube-api-access-lst2g\") pod \"controller-manager-79794f6f4c-zf854\" (UID: \"d57a7c0e-c79a-43da-a27b-f4b7f991908e\") " pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.413802 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.460015 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4"] Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.467279 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-6mgj4"] Nov 28 15:32:50 crc kubenswrapper[4805]: I1128 15:32:50.842257 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79794f6f4c-zf854"] Nov 28 15:32:50 crc kubenswrapper[4805]: W1128 15:32:50.850930 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd57a7c0e_c79a_43da_a27b_f4b7f991908e.slice/crio-5e8e4ae422582f4b41a138aeb4ecd7f99445c68069bbf5f6cd286081d522b08a WatchSource:0}: Error finding container 5e8e4ae422582f4b41a138aeb4ecd7f99445c68069bbf5f6cd286081d522b08a: Status 404 returned error can't find the container with id 5e8e4ae422582f4b41a138aeb4ecd7f99445c68069bbf5f6cd286081d522b08a Nov 28 15:32:51 crc kubenswrapper[4805]: I1128 15:32:51.136454 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" event={"ID":"d57a7c0e-c79a-43da-a27b-f4b7f991908e","Type":"ContainerStarted","Data":"1752f9ee25604d0e46de8453ba9fee68d497ac7cbdc145c4c7f3df78795f27cd"} Nov 28 15:32:51 crc kubenswrapper[4805]: I1128 15:32:51.136502 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" event={"ID":"d57a7c0e-c79a-43da-a27b-f4b7f991908e","Type":"ContainerStarted","Data":"5e8e4ae422582f4b41a138aeb4ecd7f99445c68069bbf5f6cd286081d522b08a"} Nov 28 15:32:51 crc kubenswrapper[4805]: I1128 15:32:51.137554 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:32:51 crc kubenswrapper[4805]: I1128 15:32:51.140553 4805 patch_prober.go:28] interesting pod/controller-manager-79794f6f4c-zf854 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.70:8443/healthz\": dial tcp 10.217.0.70:8443: connect: connection refused" start-of-body= Nov 28 15:32:51 crc kubenswrapper[4805]: I1128 15:32:51.140599 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" podUID="d57a7c0e-c79a-43da-a27b-f4b7f991908e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.70:8443/healthz\": dial tcp 10.217.0.70:8443: connect: connection refused" Nov 28 15:32:51 crc kubenswrapper[4805]: I1128 15:32:51.159118 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" podStartSLOduration=6.159102247 podStartE2EDuration="6.159102247s" podCreationTimestamp="2025-11-28 15:32:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:32:51.156405687 +0000 UTC m=+398.206197008" watchObservedRunningTime="2025-11-28 15:32:51.159102247 +0000 UTC m=+398.208893558" Nov 28 15:32:51 crc kubenswrapper[4805]: I1128 15:32:51.213196 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf109684-739e-4bdd-94bb-8ba8d975b187" path="/var/lib/kubelet/pods/bf109684-739e-4bdd-94bb-8ba8d975b187/volumes" Nov 28 15:32:52 crc kubenswrapper[4805]: I1128 15:32:52.147246 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-79794f6f4c-zf854" Nov 28 15:33:11 crc kubenswrapper[4805]: I1128 15:33:11.060681 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:33:11 crc kubenswrapper[4805]: I1128 15:33:11.061274 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:33:11 crc kubenswrapper[4805]: I1128 15:33:11.061338 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:33:11 crc kubenswrapper[4805]: I1128 15:33:11.061913 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"426a027662f6b8d978eced3586b5c9067232b166c02a8847de1e3b495b0e0a66"} pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:33:11 crc kubenswrapper[4805]: I1128 15:33:11.061980 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" containerID="cri-o://426a027662f6b8d978eced3586b5c9067232b166c02a8847de1e3b495b0e0a66" gracePeriod=600 Nov 28 15:33:11 crc kubenswrapper[4805]: I1128 15:33:11.274176 4805 generic.go:334] "Generic (PLEG): container finished" podID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerID="426a027662f6b8d978eced3586b5c9067232b166c02a8847de1e3b495b0e0a66" exitCode=0 Nov 28 15:33:11 crc kubenswrapper[4805]: I1128 15:33:11.274283 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerDied","Data":"426a027662f6b8d978eced3586b5c9067232b166c02a8847de1e3b495b0e0a66"} Nov 28 15:33:11 crc kubenswrapper[4805]: I1128 15:33:11.274604 4805 scope.go:117] "RemoveContainer" containerID="89a61d136e3dca13f7bc0d36288d950df8638b2a38674ff5bb645525acd14937" Nov 28 15:33:12 crc kubenswrapper[4805]: I1128 15:33:12.285215 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"f5d2617ab089b076ff4e5a31b53f92c3669863973a1ca9f0082ba25d82430598"} Nov 28 15:33:14 crc kubenswrapper[4805]: I1128 15:33:14.308991 4805 scope.go:117] "RemoveContainer" containerID="40e38108ff0e2151e918674138c231017d74f2d73d961a7619e88d7d7c813781" Nov 28 15:33:14 crc kubenswrapper[4805]: I1128 15:33:14.327892 4805 trace.go:236] Trace[1432699072]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-bznbp" (28-Nov-2025 15:33:13.162) (total time: 1165ms): Nov 28 15:33:14 crc kubenswrapper[4805]: Trace[1432699072]: [1.165783993s] [1.165783993s] END Nov 28 15:33:14 crc kubenswrapper[4805]: I1128 15:33:14.344184 4805 scope.go:117] "RemoveContainer" containerID="308d342b031ad03f8f812c9456b611a23305c78adb84b613a8b7a26879dcf3e1" Nov 28 15:33:14 crc kubenswrapper[4805]: I1128 15:33:14.358181 4805 scope.go:117] "RemoveContainer" containerID="fdb4a5f517074c037c101d6b6b282b3b5ebddbca5dd0e1b7375d405dec443a26" Nov 28 15:33:14 crc kubenswrapper[4805]: I1128 15:33:14.377236 4805 scope.go:117] "RemoveContainer" containerID="131268623e9a406b2b8aea7f338b0899c78d496707785da999618e2752b06c29" Nov 28 15:33:14 crc kubenswrapper[4805]: I1128 15:33:14.399809 4805 scope.go:117] "RemoveContainer" containerID="21a9046a721762d534223c14fbe19684c24ff997ed48e2810a21fe9125799f78" Nov 28 15:33:14 crc kubenswrapper[4805]: I1128 15:33:14.426307 4805 scope.go:117] "RemoveContainer" containerID="475d6be789edb4769330dbd9dcca45fa229b3c01a53b620de7e7c20f4c1e7826" Nov 28 15:33:14 crc kubenswrapper[4805]: I1128 15:33:14.614486 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" podUID="c0edd21b-3a3d-41c4-8e8b-af747bd69575" containerName="registry" containerID="cri-o://85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878" gracePeriod=30 Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.047695 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.221666 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgvb8\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-kube-api-access-pgvb8\") pod \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.221748 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-tls\") pod \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.221934 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.222069 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-certificates\") pod \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.222143 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0edd21b-3a3d-41c4-8e8b-af747bd69575-ca-trust-extracted\") pod \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.222198 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0edd21b-3a3d-41c4-8e8b-af747bd69575-installation-pull-secrets\") pod \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.222234 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-trusted-ca\") pod \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.222284 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-bound-sa-token\") pod \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\" (UID: \"c0edd21b-3a3d-41c4-8e8b-af747bd69575\") " Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.223047 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c0edd21b-3a3d-41c4-8e8b-af747bd69575" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.224160 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c0edd21b-3a3d-41c4-8e8b-af747bd69575" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.228567 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c0edd21b-3a3d-41c4-8e8b-af747bd69575" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.228892 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-kube-api-access-pgvb8" (OuterVolumeSpecName: "kube-api-access-pgvb8") pod "c0edd21b-3a3d-41c4-8e8b-af747bd69575" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575"). InnerVolumeSpecName "kube-api-access-pgvb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.230122 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0edd21b-3a3d-41c4-8e8b-af747bd69575-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c0edd21b-3a3d-41c4-8e8b-af747bd69575" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.232884 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c0edd21b-3a3d-41c4-8e8b-af747bd69575" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.234682 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c0edd21b-3a3d-41c4-8e8b-af747bd69575" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.245080 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0edd21b-3a3d-41c4-8e8b-af747bd69575-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c0edd21b-3a3d-41c4-8e8b-af747bd69575" (UID: "c0edd21b-3a3d-41c4-8e8b-af747bd69575"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.317627 4805 generic.go:334] "Generic (PLEG): container finished" podID="c0edd21b-3a3d-41c4-8e8b-af747bd69575" containerID="85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878" exitCode=0 Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.317691 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" event={"ID":"c0edd21b-3a3d-41c4-8e8b-af747bd69575","Type":"ContainerDied","Data":"85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878"} Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.317721 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.317782 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c5vvq" event={"ID":"c0edd21b-3a3d-41c4-8e8b-af747bd69575","Type":"ContainerDied","Data":"48be2fe808df3375cd359af318b0f4d9eda885373e34728ac03484f17b7cb1d5"} Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.317816 4805 scope.go:117] "RemoveContainer" containerID="85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.322932 4805 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.322959 4805 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0edd21b-3a3d-41c4-8e8b-af747bd69575-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.322971 4805 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0edd21b-3a3d-41c4-8e8b-af747bd69575-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.322981 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0edd21b-3a3d-41c4-8e8b-af747bd69575-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.322989 4805 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.322996 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgvb8\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-kube-api-access-pgvb8\") on node \"crc\" DevicePath \"\"" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.323005 4805 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0edd21b-3a3d-41c4-8e8b-af747bd69575-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.334942 4805 scope.go:117] "RemoveContainer" containerID="85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878" Nov 28 15:33:15 crc kubenswrapper[4805]: E1128 15:33:15.335353 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878\": container with ID starting with 85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878 not found: ID does not exist" containerID="85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.335408 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878"} err="failed to get container status \"85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878\": rpc error: code = NotFound desc = could not find container \"85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878\": container with ID starting with 85ecfd0beb22ff4d233f665cef48505b48c7790bea7323dc932d59a6ab2c5878 not found: ID does not exist" Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.349746 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5vvq"] Nov 28 15:33:15 crc kubenswrapper[4805]: I1128 15:33:15.353267 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5vvq"] Nov 28 15:33:17 crc kubenswrapper[4805]: I1128 15:33:17.212202 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0edd21b-3a3d-41c4-8e8b-af747bd69575" path="/var/lib/kubelet/pods/c0edd21b-3a3d-41c4-8e8b-af747bd69575/volumes" Nov 28 15:35:11 crc kubenswrapper[4805]: I1128 15:35:11.060344 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:35:11 crc kubenswrapper[4805]: I1128 15:35:11.061225 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:35:41 crc kubenswrapper[4805]: I1128 15:35:41.060339 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:35:41 crc kubenswrapper[4805]: I1128 15:35:41.061120 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:36:11 crc kubenswrapper[4805]: I1128 15:36:11.059945 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:36:11 crc kubenswrapper[4805]: I1128 15:36:11.061468 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:36:11 crc kubenswrapper[4805]: I1128 15:36:11.061545 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:36:11 crc kubenswrapper[4805]: I1128 15:36:11.062266 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5d2617ab089b076ff4e5a31b53f92c3669863973a1ca9f0082ba25d82430598"} pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:36:11 crc kubenswrapper[4805]: I1128 15:36:11.062389 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" containerID="cri-o://f5d2617ab089b076ff4e5a31b53f92c3669863973a1ca9f0082ba25d82430598" gracePeriod=600 Nov 28 15:36:11 crc kubenswrapper[4805]: I1128 15:36:11.438481 4805 generic.go:334] "Generic (PLEG): container finished" podID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerID="f5d2617ab089b076ff4e5a31b53f92c3669863973a1ca9f0082ba25d82430598" exitCode=0 Nov 28 15:36:11 crc kubenswrapper[4805]: I1128 15:36:11.438545 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerDied","Data":"f5d2617ab089b076ff4e5a31b53f92c3669863973a1ca9f0082ba25d82430598"} Nov 28 15:36:11 crc kubenswrapper[4805]: I1128 15:36:11.439132 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"1059ec0f76219a39add1911f16614ef92240c8775a67469a3bc7df93d8da629e"} Nov 28 15:36:11 crc kubenswrapper[4805]: I1128 15:36:11.439202 4805 scope.go:117] "RemoveContainer" containerID="426a027662f6b8d978eced3586b5c9067232b166c02a8847de1e3b495b0e0a66" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.257146 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-kc58f"] Nov 28 15:37:53 crc kubenswrapper[4805]: E1128 15:37:53.258197 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0edd21b-3a3d-41c4-8e8b-af747bd69575" containerName="registry" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.258219 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0edd21b-3a3d-41c4-8e8b-af747bd69575" containerName="registry" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.258524 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0edd21b-3a3d-41c4-8e8b-af747bd69575" containerName="registry" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.272851 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-kc58f"] Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.273038 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.276397 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.276808 4805 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-xjksz" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.279236 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.279308 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.348870 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/196e3652-0148-44f3-a306-2a4cf1a183f1-node-mnt\") pod \"crc-storage-crc-kc58f\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.349294 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/196e3652-0148-44f3-a306-2a4cf1a183f1-crc-storage\") pod \"crc-storage-crc-kc58f\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.349349 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrcsn\" (UniqueName: \"kubernetes.io/projected/196e3652-0148-44f3-a306-2a4cf1a183f1-kube-api-access-vrcsn\") pod \"crc-storage-crc-kc58f\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.451130 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/196e3652-0148-44f3-a306-2a4cf1a183f1-node-mnt\") pod \"crc-storage-crc-kc58f\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.451330 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/196e3652-0148-44f3-a306-2a4cf1a183f1-crc-storage\") pod \"crc-storage-crc-kc58f\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.451402 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrcsn\" (UniqueName: \"kubernetes.io/projected/196e3652-0148-44f3-a306-2a4cf1a183f1-kube-api-access-vrcsn\") pod \"crc-storage-crc-kc58f\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.451699 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/196e3652-0148-44f3-a306-2a4cf1a183f1-node-mnt\") pod \"crc-storage-crc-kc58f\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.452748 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/196e3652-0148-44f3-a306-2a4cf1a183f1-crc-storage\") pod \"crc-storage-crc-kc58f\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.480213 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrcsn\" (UniqueName: \"kubernetes.io/projected/196e3652-0148-44f3-a306-2a4cf1a183f1-kube-api-access-vrcsn\") pod \"crc-storage-crc-kc58f\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.596243 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.901301 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-kc58f"] Nov 28 15:37:53 crc kubenswrapper[4805]: I1128 15:37:53.909460 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:37:54 crc kubenswrapper[4805]: I1128 15:37:54.136963 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-kc58f" event={"ID":"196e3652-0148-44f3-a306-2a4cf1a183f1","Type":"ContainerStarted","Data":"7030eba7c945fb67617ca1040485176d349120b768bcf4af7b219faae84c6843"} Nov 28 15:37:56 crc kubenswrapper[4805]: I1128 15:37:56.153973 4805 generic.go:334] "Generic (PLEG): container finished" podID="196e3652-0148-44f3-a306-2a4cf1a183f1" containerID="3a56fe8f4f22bd8bc17f569fd3cbcd5fe67b479d23018f65e658d510bf19a629" exitCode=0 Nov 28 15:37:56 crc kubenswrapper[4805]: I1128 15:37:56.154087 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-kc58f" event={"ID":"196e3652-0148-44f3-a306-2a4cf1a183f1","Type":"ContainerDied","Data":"3a56fe8f4f22bd8bc17f569fd3cbcd5fe67b479d23018f65e658d510bf19a629"} Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.481839 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.519510 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrcsn\" (UniqueName: \"kubernetes.io/projected/196e3652-0148-44f3-a306-2a4cf1a183f1-kube-api-access-vrcsn\") pod \"196e3652-0148-44f3-a306-2a4cf1a183f1\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.519600 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/196e3652-0148-44f3-a306-2a4cf1a183f1-node-mnt\") pod \"196e3652-0148-44f3-a306-2a4cf1a183f1\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.519661 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/196e3652-0148-44f3-a306-2a4cf1a183f1-crc-storage\") pod \"196e3652-0148-44f3-a306-2a4cf1a183f1\" (UID: \"196e3652-0148-44f3-a306-2a4cf1a183f1\") " Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.519754 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/196e3652-0148-44f3-a306-2a4cf1a183f1-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "196e3652-0148-44f3-a306-2a4cf1a183f1" (UID: "196e3652-0148-44f3-a306-2a4cf1a183f1"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.519954 4805 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/196e3652-0148-44f3-a306-2a4cf1a183f1-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.524681 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/196e3652-0148-44f3-a306-2a4cf1a183f1-kube-api-access-vrcsn" (OuterVolumeSpecName: "kube-api-access-vrcsn") pod "196e3652-0148-44f3-a306-2a4cf1a183f1" (UID: "196e3652-0148-44f3-a306-2a4cf1a183f1"). InnerVolumeSpecName "kube-api-access-vrcsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.534983 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/196e3652-0148-44f3-a306-2a4cf1a183f1-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "196e3652-0148-44f3-a306-2a4cf1a183f1" (UID: "196e3652-0148-44f3-a306-2a4cf1a183f1"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.621003 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrcsn\" (UniqueName: \"kubernetes.io/projected/196e3652-0148-44f3-a306-2a4cf1a183f1-kube-api-access-vrcsn\") on node \"crc\" DevicePath \"\"" Nov 28 15:37:57 crc kubenswrapper[4805]: I1128 15:37:57.621036 4805 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/196e3652-0148-44f3-a306-2a4cf1a183f1-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 28 15:37:58 crc kubenswrapper[4805]: I1128 15:37:58.172573 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-kc58f" event={"ID":"196e3652-0148-44f3-a306-2a4cf1a183f1","Type":"ContainerDied","Data":"7030eba7c945fb67617ca1040485176d349120b768bcf4af7b219faae84c6843"} Nov 28 15:37:58 crc kubenswrapper[4805]: I1128 15:37:58.172631 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7030eba7c945fb67617ca1040485176d349120b768bcf4af7b219faae84c6843" Nov 28 15:37:58 crc kubenswrapper[4805]: I1128 15:37:58.172656 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kc58f" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.355890 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq"] Nov 28 15:38:05 crc kubenswrapper[4805]: E1128 15:38:05.356748 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="196e3652-0148-44f3-a306-2a4cf1a183f1" containerName="storage" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.356761 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="196e3652-0148-44f3-a306-2a4cf1a183f1" containerName="storage" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.356857 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="196e3652-0148-44f3-a306-2a4cf1a183f1" containerName="storage" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.357567 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.359666 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.368026 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq"] Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.484663 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.484837 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.484881 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55clq\" (UniqueName: \"kubernetes.io/projected/e35268b6-e3cf-4e6b-9248-44904ec877cc-kube-api-access-55clq\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.586791 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.586891 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55clq\" (UniqueName: \"kubernetes.io/projected/e35268b6-e3cf-4e6b-9248-44904ec877cc-kube-api-access-55clq\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.586965 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.588048 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.588061 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.626170 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55clq\" (UniqueName: \"kubernetes.io/projected/e35268b6-e3cf-4e6b-9248-44904ec877cc-kube-api-access-55clq\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.695743 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:05 crc kubenswrapper[4805]: I1128 15:38:05.978896 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq"] Nov 28 15:38:06 crc kubenswrapper[4805]: I1128 15:38:06.230198 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" event={"ID":"e35268b6-e3cf-4e6b-9248-44904ec877cc","Type":"ContainerStarted","Data":"4539a8e5badee2e5e3ee222246a2da56a4269d0d5e96f294ddb7bb064ecb6ea4"} Nov 28 15:38:06 crc kubenswrapper[4805]: I1128 15:38:06.230684 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" event={"ID":"e35268b6-e3cf-4e6b-9248-44904ec877cc","Type":"ContainerStarted","Data":"4560f35c8dfad24a2750c37fa62b72536dd46f88daa86c829a29646c62d1a3d4"} Nov 28 15:38:07 crc kubenswrapper[4805]: I1128 15:38:07.242603 4805 generic.go:334] "Generic (PLEG): container finished" podID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerID="4539a8e5badee2e5e3ee222246a2da56a4269d0d5e96f294ddb7bb064ecb6ea4" exitCode=0 Nov 28 15:38:07 crc kubenswrapper[4805]: I1128 15:38:07.242657 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" event={"ID":"e35268b6-e3cf-4e6b-9248-44904ec877cc","Type":"ContainerDied","Data":"4539a8e5badee2e5e3ee222246a2da56a4269d0d5e96f294ddb7bb064ecb6ea4"} Nov 28 15:38:09 crc kubenswrapper[4805]: I1128 15:38:09.257906 4805 generic.go:334] "Generic (PLEG): container finished" podID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerID="c2c341dab13416bbd83ebd70a59b0e828341baad61ccf4787c2766c7dc0dacd6" exitCode=0 Nov 28 15:38:09 crc kubenswrapper[4805]: I1128 15:38:09.258032 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" event={"ID":"e35268b6-e3cf-4e6b-9248-44904ec877cc","Type":"ContainerDied","Data":"c2c341dab13416bbd83ebd70a59b0e828341baad61ccf4787c2766c7dc0dacd6"} Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.270322 4805 generic.go:334] "Generic (PLEG): container finished" podID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerID="b74f961f1b0e27ea9dec685d3ac0b0d606fdbe4e7396e646f7940b07ea0b4527" exitCode=0 Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.270466 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" event={"ID":"e35268b6-e3cf-4e6b-9248-44904ec877cc","Type":"ContainerDied","Data":"b74f961f1b0e27ea9dec685d3ac0b0d606fdbe4e7396e646f7940b07ea0b4527"} Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.546016 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6v4kb"] Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.546616 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovn-controller" containerID="cri-o://1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8" gracePeriod=30 Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.546662 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="nbdb" containerID="cri-o://1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233" gracePeriod=30 Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.546674 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd" gracePeriod=30 Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.546701 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="northd" containerID="cri-o://ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a" gracePeriod=30 Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.546699 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kube-rbac-proxy-node" containerID="cri-o://206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8" gracePeriod=30 Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.546803 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="sbdb" containerID="cri-o://8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666" gracePeriod=30 Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.546719 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovn-acl-logging" containerID="cri-o://f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d" gracePeriod=30 Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.604992 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" containerID="cri-o://500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc" gracePeriod=30 Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.907710 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/3.log" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.910101 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovn-acl-logging/0.log" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.910787 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovn-controller/0.log" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.911338 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983173 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dnx8r"] Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983449 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="nbdb" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983464 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="nbdb" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983474 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="northd" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983482 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="northd" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983497 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovn-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983506 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovn-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983522 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983531 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983541 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983549 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983558 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kube-rbac-proxy-node" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983566 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kube-rbac-proxy-node" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983577 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="sbdb" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983585 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="sbdb" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983601 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovn-acl-logging" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983609 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovn-acl-logging" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983620 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983628 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983643 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983654 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983670 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983681 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983698 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kubecfg-setup" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983708 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kubecfg-setup" Nov 28 15:38:10 crc kubenswrapper[4805]: E1128 15:38:10.983727 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983738 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983893 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovn-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983909 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983920 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983960 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="sbdb" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983976 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="kube-rbac-proxy-node" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983988 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovn-acl-logging" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.983999 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="northd" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.984011 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.984022 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.984031 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="nbdb" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.984246 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.984509 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerName="ovnkube-controller" Nov 28 15:38:10 crc kubenswrapper[4805]: I1128 15:38:10.987836 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.059808 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.059880 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.074779 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-bin\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.074819 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-openvswitch\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.075347 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.085614 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-ovn-kubernetes\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.085726 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-netns\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.085790 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-script-lib\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.085799 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.085829 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-log-socket\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.085838 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.085886 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-var-lib-openvswitch\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.085927 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-systemd\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.085966 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-node-log\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086019 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-config\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086060 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-netd\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086090 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-slash\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086137 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086193 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpjph\" (UniqueName: \"kubernetes.io/projected/412627f3-6ef3-401d-b7ae-a839d70a46b3-kube-api-access-fpjph\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086266 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-kubelet\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086314 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-systemd-units\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086402 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086452 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-log-socket" (OuterVolumeSpecName: "log-socket") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086468 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086463 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-env-overrides\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086497 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086545 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-slash" (OuterVolumeSpecName: "host-slash") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086570 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-node-log" (OuterVolumeSpecName: "node-log") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.086939 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.087091 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.087146 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.087205 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-ovn\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.087244 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.087478 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.087550 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.088057 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-etc-openvswitch\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.088106 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovn-node-metrics-cert\") pod \"412627f3-6ef3-401d-b7ae-a839d70a46b3\" (UID: \"412627f3-6ef3-401d-b7ae-a839d70a46b3\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.088160 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.088336 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-etc-openvswitch\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.088459 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-ovnkube-script-lib\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.088502 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-node-log\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.088552 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkfff\" (UniqueName: \"kubernetes.io/projected/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-kube-api-access-pkfff\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.075273 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.088782 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-log-socket\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.089100 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-slash\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.089197 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-run-netns\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.089265 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-systemd-units\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.089846 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-ovnkube-config\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.089932 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-ovn-node-metrics-cert\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090273 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-run-openvswitch\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090408 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090470 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-run-ovn-kubernetes\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090527 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-run-ovn\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090561 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-cni-bin\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090591 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-kubelet\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090619 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-run-systemd\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090653 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-cni-netd\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090685 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-env-overrides\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090789 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-var-lib-openvswitch\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090948 4805 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.090993 4805 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091023 4805 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091050 4805 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091082 4805 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091108 4805 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091136 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091162 4805 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-log-socket\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091187 4805 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091213 4805 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-node-log\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091239 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091265 4805 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091291 4805 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-slash\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091318 4805 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091348 4805 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091413 4805 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.091439 4805 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/412627f3-6ef3-401d-b7ae-a839d70a46b3-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.093102 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/412627f3-6ef3-401d-b7ae-a839d70a46b3-kube-api-access-fpjph" (OuterVolumeSpecName: "kube-api-access-fpjph") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "kube-api-access-fpjph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.093892 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.103754 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "412627f3-6ef3-401d-b7ae-a839d70a46b3" (UID: "412627f3-6ef3-401d-b7ae-a839d70a46b3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.192999 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-systemd-units\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193050 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-ovnkube-config\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193084 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-ovn-node-metrics-cert\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193132 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-run-openvswitch\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193148 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-systemd-units\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193183 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193223 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-run-ovn-kubernetes\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-run-ovn\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193286 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-cni-bin\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193311 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-run-systemd\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193336 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-kubelet\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193393 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-env-overrides\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193343 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-run-openvswitch\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193451 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-cni-netd\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193417 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-cni-netd\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193525 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-run-ovn-kubernetes\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193572 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-run-ovn\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193495 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193636 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-run-systemd\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193611 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-cni-bin\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193674 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-kubelet\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193725 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-var-lib-openvswitch\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193817 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-etc-openvswitch\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193843 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-var-lib-openvswitch\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193882 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-ovnkube-script-lib\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193928 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-node-log\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.193983 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkfff\" (UniqueName: \"kubernetes.io/projected/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-kube-api-access-pkfff\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194023 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-etc-openvswitch\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194075 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-log-socket\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194133 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-node-log\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194148 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-slash\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194215 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-slash\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194229 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-run-netns\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194350 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/412627f3-6ef3-401d-b7ae-a839d70a46b3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194418 4805 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/412627f3-6ef3-401d-b7ae-a839d70a46b3-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194426 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-env-overrides\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194446 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpjph\" (UniqueName: \"kubernetes.io/projected/412627f3-6ef3-401d-b7ae-a839d70a46b3-kube-api-access-fpjph\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194437 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-ovnkube-config\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194494 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-log-socket\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194527 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-host-run-netns\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.194585 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-ovnkube-script-lib\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.198535 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-ovn-node-metrics-cert\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.217930 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkfff\" (UniqueName: \"kubernetes.io/projected/2cd6dd51-f670-4d3f-b96b-2d6fca467edd-kube-api-access-pkfff\") pod \"ovnkube-node-dnx8r\" (UID: \"2cd6dd51-f670-4d3f-b96b-2d6fca467edd\") " pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.282061 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovnkube-controller/3.log" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.288474 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovn-acl-logging/0.log" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289136 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6v4kb_412627f3-6ef3-401d-b7ae-a839d70a46b3/ovn-controller/0.log" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289644 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc" exitCode=0 Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289708 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666" exitCode=0 Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289733 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233" exitCode=0 Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289759 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a" exitCode=0 Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289777 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd" exitCode=0 Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289794 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8" exitCode=0 Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289812 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d" exitCode=143 Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289831 4805 generic.go:334] "Generic (PLEG): container finished" podID="412627f3-6ef3-401d-b7ae-a839d70a46b3" containerID="1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8" exitCode=143 Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289726 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289918 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289875 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289961 4805 scope.go:117] "RemoveContainer" containerID="500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.289945 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290141 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290186 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290212 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290233 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290252 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290264 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290276 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290287 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290300 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290312 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290323 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290334 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290350 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290393 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290408 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290419 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290431 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290442 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290453 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290465 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290476 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290487 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290497 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290512 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290529 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290541 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290553 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290565 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290577 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290587 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290598 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290609 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290619 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290630 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290657 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6v4kb" event={"ID":"412627f3-6ef3-401d-b7ae-a839d70a46b3","Type":"ContainerDied","Data":"52d2a7adc518626b5899bd59433b5ecaf2b00cfb66b7e4c01657d37b37955b48"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290674 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290687 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290697 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290708 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290719 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290730 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290742 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290753 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290765 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.290776 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.292640 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/2.log" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.293647 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/1.log" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.293744 4805 generic.go:334] "Generic (PLEG): container finished" podID="9d0031c5-0433-419e-9363-66eb48341a68" containerID="7b9dc44d838a23520473b186116f6a70691671146dfb822f52de1795381db630" exitCode=2 Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.294137 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fv2dw" event={"ID":"9d0031c5-0433-419e-9363-66eb48341a68","Type":"ContainerDied","Data":"7b9dc44d838a23520473b186116f6a70691671146dfb822f52de1795381db630"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.294195 4805 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55"} Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.295353 4805 scope.go:117] "RemoveContainer" containerID="7b9dc44d838a23520473b186116f6a70691671146dfb822f52de1795381db630" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.296020 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-fv2dw_openshift-multus(9d0031c5-0433-419e-9363-66eb48341a68)\"" pod="openshift-multus/multus-fv2dw" podUID="9d0031c5-0433-419e-9363-66eb48341a68" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.306874 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.330192 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6v4kb"] Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.332136 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.334463 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6v4kb"] Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.365597 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.375044 4805 scope.go:117] "RemoveContainer" containerID="8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666" Nov 28 15:38:11 crc kubenswrapper[4805]: W1128 15:38:11.391604 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cd6dd51_f670_4d3f_b96b_2d6fca467edd.slice/crio-c0eb3bc839f010a5c96a15d6538d1292c2d0356ce1776d201853ebdf8db046ff WatchSource:0}: Error finding container c0eb3bc839f010a5c96a15d6538d1292c2d0356ce1776d201853ebdf8db046ff: Status 404 returned error can't find the container with id c0eb3bc839f010a5c96a15d6538d1292c2d0356ce1776d201853ebdf8db046ff Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.397182 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-util\") pod \"e35268b6-e3cf-4e6b-9248-44904ec877cc\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.397248 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55clq\" (UniqueName: \"kubernetes.io/projected/e35268b6-e3cf-4e6b-9248-44904ec877cc-kube-api-access-55clq\") pod \"e35268b6-e3cf-4e6b-9248-44904ec877cc\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.397283 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-bundle\") pod \"e35268b6-e3cf-4e6b-9248-44904ec877cc\" (UID: \"e35268b6-e3cf-4e6b-9248-44904ec877cc\") " Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.397862 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-bundle" (OuterVolumeSpecName: "bundle") pod "e35268b6-e3cf-4e6b-9248-44904ec877cc" (UID: "e35268b6-e3cf-4e6b-9248-44904ec877cc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.399840 4805 scope.go:117] "RemoveContainer" containerID="1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.402669 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e35268b6-e3cf-4e6b-9248-44904ec877cc-kube-api-access-55clq" (OuterVolumeSpecName: "kube-api-access-55clq") pod "e35268b6-e3cf-4e6b-9248-44904ec877cc" (UID: "e35268b6-e3cf-4e6b-9248-44904ec877cc"). InnerVolumeSpecName "kube-api-access-55clq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.411687 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-util" (OuterVolumeSpecName: "util") pod "e35268b6-e3cf-4e6b-9248-44904ec877cc" (UID: "e35268b6-e3cf-4e6b-9248-44904ec877cc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.457588 4805 scope.go:117] "RemoveContainer" containerID="ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.470555 4805 scope.go:117] "RemoveContainer" containerID="0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.484093 4805 scope.go:117] "RemoveContainer" containerID="206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.499216 4805 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-util\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.499241 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55clq\" (UniqueName: \"kubernetes.io/projected/e35268b6-e3cf-4e6b-9248-44904ec877cc-kube-api-access-55clq\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.499253 4805 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e35268b6-e3cf-4e6b-9248-44904ec877cc-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.516620 4805 scope.go:117] "RemoveContainer" containerID="f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.535172 4805 scope.go:117] "RemoveContainer" containerID="1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.550802 4805 scope.go:117] "RemoveContainer" containerID="f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.565493 4805 scope.go:117] "RemoveContainer" containerID="500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.565807 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": container with ID starting with 500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc not found: ID does not exist" containerID="500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.565846 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} err="failed to get container status \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": rpc error: code = NotFound desc = could not find container \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": container with ID starting with 500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.565873 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.566186 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\": container with ID starting with abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b not found: ID does not exist" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.566235 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} err="failed to get container status \"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\": rpc error: code = NotFound desc = could not find container \"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\": container with ID starting with abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.566268 4805 scope.go:117] "RemoveContainer" containerID="8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.566575 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\": container with ID starting with 8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666 not found: ID does not exist" containerID="8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.566610 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} err="failed to get container status \"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\": rpc error: code = NotFound desc = could not find container \"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\": container with ID starting with 8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.566639 4805 scope.go:117] "RemoveContainer" containerID="1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.566917 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\": container with ID starting with 1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233 not found: ID does not exist" containerID="1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.566953 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} err="failed to get container status \"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\": rpc error: code = NotFound desc = could not find container \"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\": container with ID starting with 1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.566970 4805 scope.go:117] "RemoveContainer" containerID="ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.567457 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\": container with ID starting with ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a not found: ID does not exist" containerID="ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.567509 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} err="failed to get container status \"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\": rpc error: code = NotFound desc = could not find container \"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\": container with ID starting with ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.567526 4805 scope.go:117] "RemoveContainer" containerID="0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.567789 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\": container with ID starting with 0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd not found: ID does not exist" containerID="0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.567813 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} err="failed to get container status \"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\": rpc error: code = NotFound desc = could not find container \"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\": container with ID starting with 0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.567829 4805 scope.go:117] "RemoveContainer" containerID="206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.568078 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\": container with ID starting with 206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8 not found: ID does not exist" containerID="206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.568102 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} err="failed to get container status \"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\": rpc error: code = NotFound desc = could not find container \"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\": container with ID starting with 206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.568119 4805 scope.go:117] "RemoveContainer" containerID="f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.568380 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\": container with ID starting with f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d not found: ID does not exist" containerID="f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.568400 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} err="failed to get container status \"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\": rpc error: code = NotFound desc = could not find container \"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\": container with ID starting with f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.568419 4805 scope.go:117] "RemoveContainer" containerID="1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.568632 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\": container with ID starting with 1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8 not found: ID does not exist" containerID="1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.568651 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} err="failed to get container status \"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\": rpc error: code = NotFound desc = could not find container \"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\": container with ID starting with 1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.568667 4805 scope.go:117] "RemoveContainer" containerID="f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff" Nov 28 15:38:11 crc kubenswrapper[4805]: E1128 15:38:11.568897 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\": container with ID starting with f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff not found: ID does not exist" containerID="f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.568922 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff"} err="failed to get container status \"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\": rpc error: code = NotFound desc = could not find container \"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\": container with ID starting with f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.568939 4805 scope.go:117] "RemoveContainer" containerID="500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.569152 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} err="failed to get container status \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": rpc error: code = NotFound desc = could not find container \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": container with ID starting with 500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.569179 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.570164 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} err="failed to get container status \"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\": rpc error: code = NotFound desc = could not find container \"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\": container with ID starting with abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.570240 4805 scope.go:117] "RemoveContainer" containerID="8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.571068 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} err="failed to get container status \"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\": rpc error: code = NotFound desc = could not find container \"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\": container with ID starting with 8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.571097 4805 scope.go:117] "RemoveContainer" containerID="1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.571448 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} err="failed to get container status \"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\": rpc error: code = NotFound desc = could not find container \"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\": container with ID starting with 1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.571471 4805 scope.go:117] "RemoveContainer" containerID="ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.571679 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} err="failed to get container status \"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\": rpc error: code = NotFound desc = could not find container \"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\": container with ID starting with ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.571699 4805 scope.go:117] "RemoveContainer" containerID="0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.571893 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} err="failed to get container status \"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\": rpc error: code = NotFound desc = could not find container \"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\": container with ID starting with 0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.571912 4805 scope.go:117] "RemoveContainer" containerID="206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.572158 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} err="failed to get container status \"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\": rpc error: code = NotFound desc = could not find container \"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\": container with ID starting with 206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.572184 4805 scope.go:117] "RemoveContainer" containerID="f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.572494 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} err="failed to get container status \"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\": rpc error: code = NotFound desc = could not find container \"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\": container with ID starting with f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.572514 4805 scope.go:117] "RemoveContainer" containerID="1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.572736 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} err="failed to get container status \"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\": rpc error: code = NotFound desc = could not find container \"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\": container with ID starting with 1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.572757 4805 scope.go:117] "RemoveContainer" containerID="f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.572990 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff"} err="failed to get container status \"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\": rpc error: code = NotFound desc = could not find container \"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\": container with ID starting with f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.573013 4805 scope.go:117] "RemoveContainer" containerID="500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.573225 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} err="failed to get container status \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": rpc error: code = NotFound desc = could not find container \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": container with ID starting with 500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.573242 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.573523 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} err="failed to get container status \"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\": rpc error: code = NotFound desc = could not find container \"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\": container with ID starting with abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.573543 4805 scope.go:117] "RemoveContainer" containerID="8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.573780 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} err="failed to get container status \"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\": rpc error: code = NotFound desc = could not find container \"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\": container with ID starting with 8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.573797 4805 scope.go:117] "RemoveContainer" containerID="1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.574011 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} err="failed to get container status \"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\": rpc error: code = NotFound desc = could not find container \"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\": container with ID starting with 1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.574029 4805 scope.go:117] "RemoveContainer" containerID="ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.574239 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} err="failed to get container status \"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\": rpc error: code = NotFound desc = could not find container \"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\": container with ID starting with ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.574256 4805 scope.go:117] "RemoveContainer" containerID="0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.574530 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} err="failed to get container status \"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\": rpc error: code = NotFound desc = could not find container \"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\": container with ID starting with 0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.574628 4805 scope.go:117] "RemoveContainer" containerID="206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.574875 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} err="failed to get container status \"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\": rpc error: code = NotFound desc = could not find container \"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\": container with ID starting with 206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.574892 4805 scope.go:117] "RemoveContainer" containerID="f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.575065 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} err="failed to get container status \"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\": rpc error: code = NotFound desc = could not find container \"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\": container with ID starting with f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.575081 4805 scope.go:117] "RemoveContainer" containerID="1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.575332 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} err="failed to get container status \"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\": rpc error: code = NotFound desc = could not find container \"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\": container with ID starting with 1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.575349 4805 scope.go:117] "RemoveContainer" containerID="f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.575632 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff"} err="failed to get container status \"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\": rpc error: code = NotFound desc = could not find container \"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\": container with ID starting with f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.575655 4805 scope.go:117] "RemoveContainer" containerID="500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.575890 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} err="failed to get container status \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": rpc error: code = NotFound desc = could not find container \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": container with ID starting with 500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.575905 4805 scope.go:117] "RemoveContainer" containerID="abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.576204 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b"} err="failed to get container status \"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\": rpc error: code = NotFound desc = could not find container \"abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b\": container with ID starting with abd691257015f06be3f9424dcbed485764550a0148d820ffd6040eb3b6c9751b not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.576228 4805 scope.go:117] "RemoveContainer" containerID="8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.576458 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666"} err="failed to get container status \"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\": rpc error: code = NotFound desc = could not find container \"8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666\": container with ID starting with 8a8b41fb84c30be9358f447f5932c636930df9b226b37aba62675192b7394666 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.576485 4805 scope.go:117] "RemoveContainer" containerID="1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.576719 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233"} err="failed to get container status \"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\": rpc error: code = NotFound desc = could not find container \"1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233\": container with ID starting with 1d498931b6571d7ba4133ea7be6f5f95b8a98cf22541cce69d46c7d5753b8233 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.576740 4805 scope.go:117] "RemoveContainer" containerID="ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.577070 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a"} err="failed to get container status \"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\": rpc error: code = NotFound desc = could not find container \"ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a\": container with ID starting with ea2f7c581cca849066f6daecbc6b87ba9b77d142477311484bc9818b6c4af74a not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.577107 4805 scope.go:117] "RemoveContainer" containerID="0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.577436 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd"} err="failed to get container status \"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\": rpc error: code = NotFound desc = could not find container \"0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd\": container with ID starting with 0d625972d6bfa2b6e8aab8811acc0378f1272b7f784a55c07ac1d494b83471dd not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.577455 4805 scope.go:117] "RemoveContainer" containerID="206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.577726 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8"} err="failed to get container status \"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\": rpc error: code = NotFound desc = could not find container \"206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8\": container with ID starting with 206d83e3e460e0b383823cb80c9f3c1f9be9b9c45a4c118bb36ea7bd455d76e8 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.577749 4805 scope.go:117] "RemoveContainer" containerID="f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.578006 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d"} err="failed to get container status \"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\": rpc error: code = NotFound desc = could not find container \"f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d\": container with ID starting with f0e0316d963abae2dac8eebc1042d7f866526b329d6a8abb367e615e90800e8d not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.578041 4805 scope.go:117] "RemoveContainer" containerID="1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.578350 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8"} err="failed to get container status \"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\": rpc error: code = NotFound desc = could not find container \"1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8\": container with ID starting with 1db974d15cace626bd74ccb82d67f9c35ee1aaad3ae5308e40d507492d5af6e8 not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.578593 4805 scope.go:117] "RemoveContainer" containerID="f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.578870 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff"} err="failed to get container status \"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\": rpc error: code = NotFound desc = could not find container \"f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff\": container with ID starting with f38a8d4fe573501e0172587226b7bf406326a0f10d014a052955cb191434f3ff not found: ID does not exist" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.578892 4805 scope.go:117] "RemoveContainer" containerID="500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc" Nov 28 15:38:11 crc kubenswrapper[4805]: I1128 15:38:11.579114 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc"} err="failed to get container status \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": rpc error: code = NotFound desc = could not find container \"500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc\": container with ID starting with 500a96d94338c703cab777deba9c0ccacd7cb2068646b8171875559349bc2dbc not found: ID does not exist" Nov 28 15:38:12 crc kubenswrapper[4805]: I1128 15:38:12.304048 4805 generic.go:334] "Generic (PLEG): container finished" podID="2cd6dd51-f670-4d3f-b96b-2d6fca467edd" containerID="aa21b279176eb76ff300f6578ad979b6465ed659acd7661c7f51edfeb5eefff3" exitCode=0 Nov 28 15:38:12 crc kubenswrapper[4805]: I1128 15:38:12.304131 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerDied","Data":"aa21b279176eb76ff300f6578ad979b6465ed659acd7661c7f51edfeb5eefff3"} Nov 28 15:38:12 crc kubenswrapper[4805]: I1128 15:38:12.304182 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerStarted","Data":"c0eb3bc839f010a5c96a15d6538d1292c2d0356ce1776d201853ebdf8db046ff"} Nov 28 15:38:12 crc kubenswrapper[4805]: I1128 15:38:12.309651 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" event={"ID":"e35268b6-e3cf-4e6b-9248-44904ec877cc","Type":"ContainerDied","Data":"4560f35c8dfad24a2750c37fa62b72536dd46f88daa86c829a29646c62d1a3d4"} Nov 28 15:38:12 crc kubenswrapper[4805]: I1128 15:38:12.309694 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4560f35c8dfad24a2750c37fa62b72536dd46f88daa86c829a29646c62d1a3d4" Nov 28 15:38:12 crc kubenswrapper[4805]: I1128 15:38:12.309764 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq" Nov 28 15:38:13 crc kubenswrapper[4805]: I1128 15:38:13.211988 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="412627f3-6ef3-401d-b7ae-a839d70a46b3" path="/var/lib/kubelet/pods/412627f3-6ef3-401d-b7ae-a839d70a46b3/volumes" Nov 28 15:38:13 crc kubenswrapper[4805]: I1128 15:38:13.319464 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerStarted","Data":"b371548e1671cd38676e5c2da3701f08f5516db481dafd86dd805ea1e56028cb"} Nov 28 15:38:13 crc kubenswrapper[4805]: I1128 15:38:13.319521 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerStarted","Data":"2ed3862daf4cac3f9add5988bc8ea4c43ff3e0ad95395c21910fbf5c7da49ce8"} Nov 28 15:38:13 crc kubenswrapper[4805]: I1128 15:38:13.319540 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerStarted","Data":"310b9452d2296326d73b4b6d76271ea9b27935adfbac6e5e4532efea9c98864e"} Nov 28 15:38:13 crc kubenswrapper[4805]: I1128 15:38:13.319557 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerStarted","Data":"e5c986087ca91770cd071eae0add62eaa08bed2b0157e72d7a459bde7bb41b6f"} Nov 28 15:38:13 crc kubenswrapper[4805]: I1128 15:38:13.319573 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerStarted","Data":"46d54e7134cbb58fb8a52b0bf4e59f992b8e3f79f32a14d3bf0cf4c807c26483"} Nov 28 15:38:13 crc kubenswrapper[4805]: I1128 15:38:13.319589 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerStarted","Data":"d6fad32e877f85ef4dff47ca6ae916cbf24aefa81ae3cd00d9d7fd828f2c4114"} Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.512996 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47"] Nov 28 15:38:14 crc kubenswrapper[4805]: E1128 15:38:14.513195 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerName="pull" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.513205 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerName="pull" Nov 28 15:38:14 crc kubenswrapper[4805]: E1128 15:38:14.513219 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerName="util" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.513225 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerName="util" Nov 28 15:38:14 crc kubenswrapper[4805]: E1128 15:38:14.513233 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerName="extract" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.513239 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerName="extract" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.513371 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e35268b6-e3cf-4e6b-9248-44904ec877cc" containerName="extract" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.513728 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.516278 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.516502 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4fp82" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.517249 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.539019 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j5g9\" (UniqueName: \"kubernetes.io/projected/82867a57-82fc-4a21-88f5-86636310e157-kube-api-access-9j5g9\") pod \"nmstate-operator-5b5b58f5c8-j6g47\" (UID: \"82867a57-82fc-4a21-88f5-86636310e157\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.543439 4805 scope.go:117] "RemoveContainer" containerID="d572efac622b246ca027cb9cf489c92f5f81c3da8ec4ad8f18765f6f4bb80e55" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.640295 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j5g9\" (UniqueName: \"kubernetes.io/projected/82867a57-82fc-4a21-88f5-86636310e157-kube-api-access-9j5g9\") pod \"nmstate-operator-5b5b58f5c8-j6g47\" (UID: \"82867a57-82fc-4a21-88f5-86636310e157\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.663017 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j5g9\" (UniqueName: \"kubernetes.io/projected/82867a57-82fc-4a21-88f5-86636310e157-kube-api-access-9j5g9\") pod \"nmstate-operator-5b5b58f5c8-j6g47\" (UID: \"82867a57-82fc-4a21-88f5-86636310e157\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:14 crc kubenswrapper[4805]: I1128 15:38:14.828773 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:14 crc kubenswrapper[4805]: E1128 15:38:14.849636 4805 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate_82867a57-82fc-4a21-88f5-86636310e157_0(ec38d42e5dcd35bacdf6ba3292c2c2ab9f53d99cd2802967a59fd8ea9f11da87): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 15:38:14 crc kubenswrapper[4805]: E1128 15:38:14.849709 4805 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate_82867a57-82fc-4a21-88f5-86636310e157_0(ec38d42e5dcd35bacdf6ba3292c2c2ab9f53d99cd2802967a59fd8ea9f11da87): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:14 crc kubenswrapper[4805]: E1128 15:38:14.849732 4805 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate_82867a57-82fc-4a21-88f5-86636310e157_0(ec38d42e5dcd35bacdf6ba3292c2c2ab9f53d99cd2802967a59fd8ea9f11da87): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:14 crc kubenswrapper[4805]: E1128 15:38:14.849776 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate(82867a57-82fc-4a21-88f5-86636310e157)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate(82867a57-82fc-4a21-88f5-86636310e157)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate_82867a57-82fc-4a21-88f5-86636310e157_0(ec38d42e5dcd35bacdf6ba3292c2c2ab9f53d99cd2802967a59fd8ea9f11da87): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" podUID="82867a57-82fc-4a21-88f5-86636310e157" Nov 28 15:38:15 crc kubenswrapper[4805]: I1128 15:38:15.335868 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/2.log" Nov 28 15:38:16 crc kubenswrapper[4805]: I1128 15:38:16.350803 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerStarted","Data":"5ef6a2391fa64146f5d721cfa6bec7f1346506dbe4b09989d30eb307276125d4"} Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.365446 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" event={"ID":"2cd6dd51-f670-4d3f-b96b-2d6fca467edd","Type":"ContainerStarted","Data":"a2dcdcd0069ba0ab04bca804f6e1b3341b96dd59309779b30107f51b0e100e1f"} Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.366015 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.366034 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.366046 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.391203 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" podStartSLOduration=8.391185225 podStartE2EDuration="8.391185225s" podCreationTimestamp="2025-11-28 15:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:38:18.389401315 +0000 UTC m=+725.439192636" watchObservedRunningTime="2025-11-28 15:38:18.391185225 +0000 UTC m=+725.440976536" Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.395724 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.396537 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.560199 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47"] Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.560331 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:18 crc kubenswrapper[4805]: I1128 15:38:18.560734 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:18 crc kubenswrapper[4805]: E1128 15:38:18.582348 4805 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate_82867a57-82fc-4a21-88f5-86636310e157_0(3a6bd0a2ddc7679ac005ad9c0aebd1f206d42c56316279fd4e497f9cbcbab632): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 15:38:18 crc kubenswrapper[4805]: E1128 15:38:18.582462 4805 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate_82867a57-82fc-4a21-88f5-86636310e157_0(3a6bd0a2ddc7679ac005ad9c0aebd1f206d42c56316279fd4e497f9cbcbab632): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:18 crc kubenswrapper[4805]: E1128 15:38:18.582501 4805 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate_82867a57-82fc-4a21-88f5-86636310e157_0(3a6bd0a2ddc7679ac005ad9c0aebd1f206d42c56316279fd4e497f9cbcbab632): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:18 crc kubenswrapper[4805]: E1128 15:38:18.582601 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate(82867a57-82fc-4a21-88f5-86636310e157)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate(82867a57-82fc-4a21-88f5-86636310e157)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-5b5b58f5c8-j6g47_openshift-nmstate_82867a57-82fc-4a21-88f5-86636310e157_0(3a6bd0a2ddc7679ac005ad9c0aebd1f206d42c56316279fd4e497f9cbcbab632): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" podUID="82867a57-82fc-4a21-88f5-86636310e157" Nov 28 15:38:26 crc kubenswrapper[4805]: I1128 15:38:26.206795 4805 scope.go:117] "RemoveContainer" containerID="7b9dc44d838a23520473b186116f6a70691671146dfb822f52de1795381db630" Nov 28 15:38:27 crc kubenswrapper[4805]: I1128 15:38:27.433722 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fv2dw_9d0031c5-0433-419e-9363-66eb48341a68/kube-multus/2.log" Nov 28 15:38:27 crc kubenswrapper[4805]: I1128 15:38:27.434036 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fv2dw" event={"ID":"9d0031c5-0433-419e-9363-66eb48341a68","Type":"ContainerStarted","Data":"3dfad7896309d6105c5ea7dd38ff3b5fa9b8f0e92431d482b3b034403cdf663f"} Nov 28 15:38:30 crc kubenswrapper[4805]: I1128 15:38:30.203971 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:30 crc kubenswrapper[4805]: I1128 15:38:30.204749 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" Nov 28 15:38:30 crc kubenswrapper[4805]: I1128 15:38:30.414599 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47"] Nov 28 15:38:30 crc kubenswrapper[4805]: I1128 15:38:30.462942 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" event={"ID":"82867a57-82fc-4a21-88f5-86636310e157","Type":"ContainerStarted","Data":"b84f8e229e08aa4fb58f7a55f579e3cda7c77255c8e6976f961e2d57f12608d0"} Nov 28 15:38:34 crc kubenswrapper[4805]: I1128 15:38:34.498933 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" event={"ID":"82867a57-82fc-4a21-88f5-86636310e157","Type":"ContainerStarted","Data":"de899a41fdf870b514d137d35a2afdad3f5918f47a00f99d45b5fa2468c70d4f"} Nov 28 15:38:35 crc kubenswrapper[4805]: I1128 15:38:35.539551 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j6g47" podStartSLOduration=18.938713537 podStartE2EDuration="21.539526559s" podCreationTimestamp="2025-11-28 15:38:14 +0000 UTC" firstStartedPulling="2025-11-28 15:38:30.432336224 +0000 UTC m=+737.482127555" lastFinishedPulling="2025-11-28 15:38:33.033149256 +0000 UTC m=+740.082940577" observedRunningTime="2025-11-28 15:38:35.535084155 +0000 UTC m=+742.584875536" watchObservedRunningTime="2025-11-28 15:38:35.539526559 +0000 UTC m=+742.589317910" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.570753 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k"] Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.571976 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.573835 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-grfgr" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.592539 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77"] Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.593293 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.596462 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.637317 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k"] Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.640550 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77"] Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.648252 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-s4w9z"] Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.648993 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.723561 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp"] Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.724340 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.726024 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.726292 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-7fb2x" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.726775 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.732728 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp"] Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.739083 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27wnr\" (UniqueName: \"kubernetes.io/projected/d83b5a88-f387-4751-b525-b1cc7f849322-kube-api-access-27wnr\") pod \"nmstate-metrics-7f946cbc9-46m8k\" (UID: \"d83b5a88-f387-4751-b525-b1cc7f849322\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.739174 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/82d7ddd8-2fc7-4e21-a235-2559df91b09c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6hn77\" (UID: \"82d7ddd8-2fc7-4e21-a235-2559df91b09c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.739326 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6pjm\" (UniqueName: \"kubernetes.io/projected/82d7ddd8-2fc7-4e21-a235-2559df91b09c-kube-api-access-t6pjm\") pod \"nmstate-webhook-5f6d4c5ccb-6hn77\" (UID: \"82d7ddd8-2fc7-4e21-a235-2559df91b09c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843258 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f62593c7-47a5-40e2-8660-591e3cbe2ad1-ovs-socket\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843312 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6pjm\" (UniqueName: \"kubernetes.io/projected/82d7ddd8-2fc7-4e21-a235-2559df91b09c-kube-api-access-t6pjm\") pod \"nmstate-webhook-5f6d4c5ccb-6hn77\" (UID: \"82d7ddd8-2fc7-4e21-a235-2559df91b09c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843338 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27wnr\" (UniqueName: \"kubernetes.io/projected/d83b5a88-f387-4751-b525-b1cc7f849322-kube-api-access-27wnr\") pod \"nmstate-metrics-7f946cbc9-46m8k\" (UID: \"d83b5a88-f387-4751-b525-b1cc7f849322\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843382 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f62593c7-47a5-40e2-8660-591e3cbe2ad1-nmstate-lock\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843654 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5e74c9e2-3720-438a-a6df-1e9083608caa-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843737 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls7k8\" (UniqueName: \"kubernetes.io/projected/5e74c9e2-3720-438a-a6df-1e9083608caa-kube-api-access-ls7k8\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843818 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f62593c7-47a5-40e2-8660-591e3cbe2ad1-dbus-socket\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843872 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw94x\" (UniqueName: \"kubernetes.io/projected/f62593c7-47a5-40e2-8660-591e3cbe2ad1-kube-api-access-hw94x\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843910 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e74c9e2-3720-438a-a6df-1e9083608caa-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.843960 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/82d7ddd8-2fc7-4e21-a235-2559df91b09c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6hn77\" (UID: \"82d7ddd8-2fc7-4e21-a235-2559df91b09c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:36 crc kubenswrapper[4805]: E1128 15:38:36.844084 4805 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 28 15:38:36 crc kubenswrapper[4805]: E1128 15:38:36.844160 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82d7ddd8-2fc7-4e21-a235-2559df91b09c-tls-key-pair podName:82d7ddd8-2fc7-4e21-a235-2559df91b09c nodeName:}" failed. No retries permitted until 2025-11-28 15:38:37.344140737 +0000 UTC m=+744.393932048 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/82d7ddd8-2fc7-4e21-a235-2559df91b09c-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-6hn77" (UID: "82d7ddd8-2fc7-4e21-a235-2559df91b09c") : secret "openshift-nmstate-webhook" not found Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.868617 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6pjm\" (UniqueName: \"kubernetes.io/projected/82d7ddd8-2fc7-4e21-a235-2559df91b09c-kube-api-access-t6pjm\") pod \"nmstate-webhook-5f6d4c5ccb-6hn77\" (UID: \"82d7ddd8-2fc7-4e21-a235-2559df91b09c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.868838 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27wnr\" (UniqueName: \"kubernetes.io/projected/d83b5a88-f387-4751-b525-b1cc7f849322-kube-api-access-27wnr\") pod \"nmstate-metrics-7f946cbc9-46m8k\" (UID: \"d83b5a88-f387-4751-b525-b1cc7f849322\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.923549 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-858544c664-dmlp7"] Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.924176 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945187 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw94x\" (UniqueName: \"kubernetes.io/projected/f62593c7-47a5-40e2-8660-591e3cbe2ad1-kube-api-access-hw94x\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945243 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e74c9e2-3720-438a-a6df-1e9083608caa-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945314 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f62593c7-47a5-40e2-8660-591e3cbe2ad1-ovs-socket\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945385 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f62593c7-47a5-40e2-8660-591e3cbe2ad1-nmstate-lock\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945420 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5e74c9e2-3720-438a-a6df-1e9083608caa-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945449 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls7k8\" (UniqueName: \"kubernetes.io/projected/5e74c9e2-3720-438a-a6df-1e9083608caa-kube-api-access-ls7k8\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945453 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f62593c7-47a5-40e2-8660-591e3cbe2ad1-ovs-socket\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945480 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f62593c7-47a5-40e2-8660-591e3cbe2ad1-dbus-socket\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945477 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f62593c7-47a5-40e2-8660-591e3cbe2ad1-nmstate-lock\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: E1128 15:38:36.945499 4805 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945466 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k" Nov 28 15:38:36 crc kubenswrapper[4805]: E1128 15:38:36.945594 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e74c9e2-3720-438a-a6df-1e9083608caa-plugin-serving-cert podName:5e74c9e2-3720-438a-a6df-1e9083608caa nodeName:}" failed. No retries permitted until 2025-11-28 15:38:37.445574172 +0000 UTC m=+744.495365483 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/5e74c9e2-3720-438a-a6df-1e9083608caa-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-bzkzp" (UID: "5e74c9e2-3720-438a-a6df-1e9083608caa") : secret "plugin-serving-cert" not found Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.945816 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f62593c7-47a5-40e2-8660-591e3cbe2ad1-dbus-socket\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.946442 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5e74c9e2-3720-438a-a6df-1e9083608caa-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.990521 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls7k8\" (UniqueName: \"kubernetes.io/projected/5e74c9e2-3720-438a-a6df-1e9083608caa-kube-api-access-ls7k8\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:36 crc kubenswrapper[4805]: I1128 15:38:36.992316 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw94x\" (UniqueName: \"kubernetes.io/projected/f62593c7-47a5-40e2-8660-591e3cbe2ad1-kube-api-access-hw94x\") pod \"nmstate-handler-s4w9z\" (UID: \"f62593c7-47a5-40e2-8660-591e3cbe2ad1\") " pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.028492 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-858544c664-dmlp7"] Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.047278 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5pqr\" (UniqueName: \"kubernetes.io/projected/ff628502-b497-4012-bbb1-bd979a0559aa-kube-api-access-v5pqr\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.047338 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-console-config\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.047371 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-oauth-serving-cert\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.047425 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff628502-b497-4012-bbb1-bd979a0559aa-console-oauth-config\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.047444 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-service-ca\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.047463 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff628502-b497-4012-bbb1-bd979a0559aa-console-serving-cert\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.047489 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-trusted-ca-bundle\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.148708 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-console-config\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.149054 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-oauth-serving-cert\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.149122 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff628502-b497-4012-bbb1-bd979a0559aa-console-oauth-config\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.149143 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-service-ca\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.149164 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff628502-b497-4012-bbb1-bd979a0559aa-console-serving-cert\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.149192 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-trusted-ca-bundle\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.149216 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5pqr\" (UniqueName: \"kubernetes.io/projected/ff628502-b497-4012-bbb1-bd979a0559aa-kube-api-access-v5pqr\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.149350 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-console-config\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.149998 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-service-ca\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.150272 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-oauth-serving-cert\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.150678 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff628502-b497-4012-bbb1-bd979a0559aa-trusted-ca-bundle\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.153155 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff628502-b497-4012-bbb1-bd979a0559aa-console-serving-cert\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.153158 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff628502-b497-4012-bbb1-bd979a0559aa-console-oauth-config\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.163681 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5pqr\" (UniqueName: \"kubernetes.io/projected/ff628502-b497-4012-bbb1-bd979a0559aa-kube-api-access-v5pqr\") pod \"console-858544c664-dmlp7\" (UID: \"ff628502-b497-4012-bbb1-bd979a0559aa\") " pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.237509 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.275117 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:37 crc kubenswrapper[4805]: W1128 15:38:37.308334 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf62593c7_47a5_40e2_8660_591e3cbe2ad1.slice/crio-17b51759ddd18ce54a215c2dc936fc182c290358000d9290491d9b8e1696a577 WatchSource:0}: Error finding container 17b51759ddd18ce54a215c2dc936fc182c290358000d9290491d9b8e1696a577: Status 404 returned error can't find the container with id 17b51759ddd18ce54a215c2dc936fc182c290358000d9290491d9b8e1696a577 Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.356708 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/82d7ddd8-2fc7-4e21-a235-2559df91b09c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6hn77\" (UID: \"82d7ddd8-2fc7-4e21-a235-2559df91b09c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.360869 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/82d7ddd8-2fc7-4e21-a235-2559df91b09c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6hn77\" (UID: \"82d7ddd8-2fc7-4e21-a235-2559df91b09c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.414155 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k"] Nov 28 15:38:37 crc kubenswrapper[4805]: W1128 15:38:37.418550 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd83b5a88_f387_4751_b525_b1cc7f849322.slice/crio-2b25ae28e4aa236fee7fb3c450a3424187bf99aaf7ae4169912f2ac76b35cab7 WatchSource:0}: Error finding container 2b25ae28e4aa236fee7fb3c450a3424187bf99aaf7ae4169912f2ac76b35cab7: Status 404 returned error can't find the container with id 2b25ae28e4aa236fee7fb3c450a3424187bf99aaf7ae4169912f2ac76b35cab7 Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.458341 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e74c9e2-3720-438a-a6df-1e9083608caa-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.463396 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e74c9e2-3720-438a-a6df-1e9083608caa-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bzkzp\" (UID: \"5e74c9e2-3720-438a-a6df-1e9083608caa\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.519880 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-s4w9z" event={"ID":"f62593c7-47a5-40e2-8660-591e3cbe2ad1","Type":"ContainerStarted","Data":"17b51759ddd18ce54a215c2dc936fc182c290358000d9290491d9b8e1696a577"} Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.521466 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k" event={"ID":"d83b5a88-f387-4751-b525-b1cc7f849322","Type":"ContainerStarted","Data":"2b25ae28e4aa236fee7fb3c450a3424187bf99aaf7ae4169912f2ac76b35cab7"} Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.557444 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.637083 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" Nov 28 15:38:37 crc kubenswrapper[4805]: I1128 15:38:37.651208 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-858544c664-dmlp7"] Nov 28 15:38:37 crc kubenswrapper[4805]: W1128 15:38:37.668588 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff628502_b497_4012_bbb1_bd979a0559aa.slice/crio-88d33e6792bc37af7d9935c704632c7c8a582b85f05d6bd82324ee97de65a03d WatchSource:0}: Error finding container 88d33e6792bc37af7d9935c704632c7c8a582b85f05d6bd82324ee97de65a03d: Status 404 returned error can't find the container with id 88d33e6792bc37af7d9935c704632c7c8a582b85f05d6bd82324ee97de65a03d Nov 28 15:38:38 crc kubenswrapper[4805]: I1128 15:38:38.122044 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp"] Nov 28 15:38:38 crc kubenswrapper[4805]: I1128 15:38:38.302594 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77"] Nov 28 15:38:38 crc kubenswrapper[4805]: I1128 15:38:38.527333 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-858544c664-dmlp7" event={"ID":"ff628502-b497-4012-bbb1-bd979a0559aa","Type":"ContainerStarted","Data":"d4384e00eb61c835fbbec98480e3f24bd998937ef67ccd41324bcd2330230a13"} Nov 28 15:38:38 crc kubenswrapper[4805]: I1128 15:38:38.527412 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-858544c664-dmlp7" event={"ID":"ff628502-b497-4012-bbb1-bd979a0559aa","Type":"ContainerStarted","Data":"88d33e6792bc37af7d9935c704632c7c8a582b85f05d6bd82324ee97de65a03d"} Nov 28 15:38:38 crc kubenswrapper[4805]: I1128 15:38:38.530875 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" event={"ID":"5e74c9e2-3720-438a-a6df-1e9083608caa","Type":"ContainerStarted","Data":"3620b29f008e45bc66441416459fdfee02b50523764640f8f5bf83a10d1285e6"} Nov 28 15:38:38 crc kubenswrapper[4805]: I1128 15:38:38.531897 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" event={"ID":"82d7ddd8-2fc7-4e21-a235-2559df91b09c","Type":"ContainerStarted","Data":"2e0aac395e4ce83c9d3ff7c56f861672f0e73fbc86b5a20b50eba5a68c207d06"} Nov 28 15:38:38 crc kubenswrapper[4805]: I1128 15:38:38.546829 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-858544c664-dmlp7" podStartSLOduration=2.546810273 podStartE2EDuration="2.546810273s" podCreationTimestamp="2025-11-28 15:38:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:38:38.542929205 +0000 UTC m=+745.592720516" watchObservedRunningTime="2025-11-28 15:38:38.546810273 +0000 UTC m=+745.596601584" Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.023154 4805 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.059684 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.059760 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.330762 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dnx8r" Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.547183 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" event={"ID":"5e74c9e2-3720-438a-a6df-1e9083608caa","Type":"ContainerStarted","Data":"78ff836f857399aa53e0c5acfaf811c39421e65b070dbaf42824c32addf42f5f"} Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.548552 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-s4w9z" event={"ID":"f62593c7-47a5-40e2-8660-591e3cbe2ad1","Type":"ContainerStarted","Data":"5cb37e4e5ef3268eaa026a1d7ca67766c901b6ea51cba7403bac6827f41ae424"} Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.548663 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.550682 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k" event={"ID":"d83b5a88-f387-4751-b525-b1cc7f849322","Type":"ContainerStarted","Data":"209c05c904690612f6efad8b9c9f175c9af515d27b53fe5beaa87480c5dfcaa1"} Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.553684 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" event={"ID":"82d7ddd8-2fc7-4e21-a235-2559df91b09c","Type":"ContainerStarted","Data":"9cc0bfaf66cf75968fec6b4d60de3afcf312df351941db8bd1caaf4cf4ebe287"} Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.554130 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.564256 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bzkzp" podStartSLOduration=3.1206357 podStartE2EDuration="5.564237459s" podCreationTimestamp="2025-11-28 15:38:36 +0000 UTC" firstStartedPulling="2025-11-28 15:38:38.136491165 +0000 UTC m=+745.186282496" lastFinishedPulling="2025-11-28 15:38:40.580092904 +0000 UTC m=+747.629884255" observedRunningTime="2025-11-28 15:38:41.560844244 +0000 UTC m=+748.610635575" watchObservedRunningTime="2025-11-28 15:38:41.564237459 +0000 UTC m=+748.614028770" Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.587168 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" podStartSLOduration=3.319489318 podStartE2EDuration="5.587132384s" podCreationTimestamp="2025-11-28 15:38:36 +0000 UTC" firstStartedPulling="2025-11-28 15:38:38.323148016 +0000 UTC m=+745.372939367" lastFinishedPulling="2025-11-28 15:38:40.590791122 +0000 UTC m=+747.640582433" observedRunningTime="2025-11-28 15:38:41.582109075 +0000 UTC m=+748.631900386" watchObservedRunningTime="2025-11-28 15:38:41.587132384 +0000 UTC m=+748.636923695" Nov 28 15:38:41 crc kubenswrapper[4805]: I1128 15:38:41.607702 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-s4w9z" podStartSLOduration=2.319742851 podStartE2EDuration="5.607686485s" podCreationTimestamp="2025-11-28 15:38:36 +0000 UTC" firstStartedPulling="2025-11-28 15:38:37.313637417 +0000 UTC m=+744.363428728" lastFinishedPulling="2025-11-28 15:38:40.601581051 +0000 UTC m=+747.651372362" observedRunningTime="2025-11-28 15:38:41.602995465 +0000 UTC m=+748.652786776" watchObservedRunningTime="2025-11-28 15:38:41.607686485 +0000 UTC m=+748.657477796" Nov 28 15:38:44 crc kubenswrapper[4805]: I1128 15:38:44.576131 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k" event={"ID":"d83b5a88-f387-4751-b525-b1cc7f849322","Type":"ContainerStarted","Data":"f1144f71c499702f3120ddfc7867e01cba1686a656eb704d4f1330c8b256baeb"} Nov 28 15:38:44 crc kubenswrapper[4805]: I1128 15:38:44.611747 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-46m8k" podStartSLOduration=2.60074641 podStartE2EDuration="8.611714489s" podCreationTimestamp="2025-11-28 15:38:36 +0000 UTC" firstStartedPulling="2025-11-28 15:38:37.421719997 +0000 UTC m=+744.471511318" lastFinishedPulling="2025-11-28 15:38:43.432688066 +0000 UTC m=+750.482479397" observedRunningTime="2025-11-28 15:38:44.60345136 +0000 UTC m=+751.653242711" watchObservedRunningTime="2025-11-28 15:38:44.611714489 +0000 UTC m=+751.661505840" Nov 28 15:38:47 crc kubenswrapper[4805]: I1128 15:38:47.238596 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:47 crc kubenswrapper[4805]: I1128 15:38:47.238684 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:47 crc kubenswrapper[4805]: I1128 15:38:47.243815 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:47 crc kubenswrapper[4805]: I1128 15:38:47.305256 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-s4w9z" Nov 28 15:38:47 crc kubenswrapper[4805]: I1128 15:38:47.601272 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-858544c664-dmlp7" Nov 28 15:38:47 crc kubenswrapper[4805]: I1128 15:38:47.688165 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7bprj"] Nov 28 15:38:57 crc kubenswrapper[4805]: I1128 15:38:57.566710 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6hn77" Nov 28 15:39:11 crc kubenswrapper[4805]: I1128 15:39:11.060770 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:39:11 crc kubenswrapper[4805]: I1128 15:39:11.061465 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:39:11 crc kubenswrapper[4805]: I1128 15:39:11.061533 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:39:11 crc kubenswrapper[4805]: I1128 15:39:11.062483 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1059ec0f76219a39add1911f16614ef92240c8775a67469a3bc7df93d8da629e"} pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:39:11 crc kubenswrapper[4805]: I1128 15:39:11.062587 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" containerID="cri-o://1059ec0f76219a39add1911f16614ef92240c8775a67469a3bc7df93d8da629e" gracePeriod=600 Nov 28 15:39:11 crc kubenswrapper[4805]: I1128 15:39:11.762720 4805 generic.go:334] "Generic (PLEG): container finished" podID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerID="1059ec0f76219a39add1911f16614ef92240c8775a67469a3bc7df93d8da629e" exitCode=0 Nov 28 15:39:11 crc kubenswrapper[4805]: I1128 15:39:11.762754 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerDied","Data":"1059ec0f76219a39add1911f16614ef92240c8775a67469a3bc7df93d8da629e"} Nov 28 15:39:11 crc kubenswrapper[4805]: I1128 15:39:11.763142 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"9401a6879e1fcc43d0b27657309da77bf635d82861975cc53606a87548ec258e"} Nov 28 15:39:11 crc kubenswrapper[4805]: I1128 15:39:11.763165 4805 scope.go:117] "RemoveContainer" containerID="f5d2617ab089b076ff4e5a31b53f92c3669863973a1ca9f0082ba25d82430598" Nov 28 15:39:12 crc kubenswrapper[4805]: I1128 15:39:12.745130 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-7bprj" podUID="69b46aab-afdf-4551-b0b4-b8d4f8f97831" containerName="console" containerID="cri-o://ff69a3cd254e168cb937fb26fabceb62c0b5a467cb11b0339d2fc49212321969" gracePeriod=15 Nov 28 15:39:12 crc kubenswrapper[4805]: E1128 15:39:12.970716 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69b46aab_afdf_4551_b0b4_b8d4f8f97831.slice/crio-conmon-ff69a3cd254e168cb937fb26fabceb62c0b5a467cb11b0339d2fc49212321969.scope\": RecentStats: unable to find data in memory cache]" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.293726 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95"] Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.296635 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.306311 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95"] Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.312837 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.464814 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpr87\" (UniqueName: \"kubernetes.io/projected/72ae900f-55af-457e-9785-8d671059454d-kube-api-access-zpr87\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.465051 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.465125 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.566782 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.566874 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpr87\" (UniqueName: \"kubernetes.io/projected/72ae900f-55af-457e-9785-8d671059454d-kube-api-access-zpr87\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.567049 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.567573 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.567740 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.593495 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpr87\" (UniqueName: \"kubernetes.io/projected/72ae900f-55af-457e-9785-8d671059454d-kube-api-access-zpr87\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.637506 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.645109 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.782324 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7bprj_69b46aab-afdf-4551-b0b4-b8d4f8f97831/console/0.log" Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.782663 4805 generic.go:334] "Generic (PLEG): container finished" podID="69b46aab-afdf-4551-b0b4-b8d4f8f97831" containerID="ff69a3cd254e168cb937fb26fabceb62c0b5a467cb11b0339d2fc49212321969" exitCode=2 Nov 28 15:39:13 crc kubenswrapper[4805]: I1128 15:39:13.782698 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7bprj" event={"ID":"69b46aab-afdf-4551-b0b4-b8d4f8f97831","Type":"ContainerDied","Data":"ff69a3cd254e168cb937fb26fabceb62c0b5a467cb11b0339d2fc49212321969"} Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.095654 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95"] Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.261404 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7bprj_69b46aab-afdf-4551-b0b4-b8d4f8f97831/console/0.log" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.261470 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.377975 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-trusted-ca-bundle\") pod \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.379115 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjzxd\" (UniqueName: \"kubernetes.io/projected/69b46aab-afdf-4551-b0b4-b8d4f8f97831-kube-api-access-vjzxd\") pod \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.379039 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "69b46aab-afdf-4551-b0b4-b8d4f8f97831" (UID: "69b46aab-afdf-4551-b0b4-b8d4f8f97831"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.379988 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-oauth-serving-cert\") pod \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.380114 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-service-ca\") pod \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.380155 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-oauth-config\") pod \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.380193 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-serving-cert\") pod \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.380222 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-config\") pod \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\" (UID: \"69b46aab-afdf-4551-b0b4-b8d4f8f97831\") " Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.380644 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "69b46aab-afdf-4551-b0b4-b8d4f8f97831" (UID: "69b46aab-afdf-4551-b0b4-b8d4f8f97831"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.380811 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.380833 4805 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.380866 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-service-ca" (OuterVolumeSpecName: "service-ca") pod "69b46aab-afdf-4551-b0b4-b8d4f8f97831" (UID: "69b46aab-afdf-4551-b0b4-b8d4f8f97831"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.381119 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-config" (OuterVolumeSpecName: "console-config") pod "69b46aab-afdf-4551-b0b4-b8d4f8f97831" (UID: "69b46aab-afdf-4551-b0b4-b8d4f8f97831"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.400617 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "69b46aab-afdf-4551-b0b4-b8d4f8f97831" (UID: "69b46aab-afdf-4551-b0b4-b8d4f8f97831"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.400726 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69b46aab-afdf-4551-b0b4-b8d4f8f97831-kube-api-access-vjzxd" (OuterVolumeSpecName: "kube-api-access-vjzxd") pod "69b46aab-afdf-4551-b0b4-b8d4f8f97831" (UID: "69b46aab-afdf-4551-b0b4-b8d4f8f97831"). InnerVolumeSpecName "kube-api-access-vjzxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.402503 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "69b46aab-afdf-4551-b0b4-b8d4f8f97831" (UID: "69b46aab-afdf-4551-b0b4-b8d4f8f97831"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.481792 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.481832 4805 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.481853 4805 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.481869 4805 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69b46aab-afdf-4551-b0b4-b8d4f8f97831-console-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.481887 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjzxd\" (UniqueName: \"kubernetes.io/projected/69b46aab-afdf-4551-b0b4-b8d4f8f97831-kube-api-access-vjzxd\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.598861 4805 scope.go:117] "RemoveContainer" containerID="ff69a3cd254e168cb937fb26fabceb62c0b5a467cb11b0339d2fc49212321969" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.790015 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7bprj" event={"ID":"69b46aab-afdf-4551-b0b4-b8d4f8f97831","Type":"ContainerDied","Data":"c11e8a56bf01975264f72042e359a8966be236c441b3486b093aeef107642284"} Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.792149 4805 generic.go:334] "Generic (PLEG): container finished" podID="72ae900f-55af-457e-9785-8d671059454d" containerID="3875f5211db3b279a179e9c1af4805455d2aace110ff7f7134ec1b4b6250616d" exitCode=0 Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.792226 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7bprj" Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.792230 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" event={"ID":"72ae900f-55af-457e-9785-8d671059454d","Type":"ContainerDied","Data":"3875f5211db3b279a179e9c1af4805455d2aace110ff7f7134ec1b4b6250616d"} Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.792265 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" event={"ID":"72ae900f-55af-457e-9785-8d671059454d","Type":"ContainerStarted","Data":"759897d352b2b42ddf24a9a4ea5144867e8aa2ecb1266bf7bc46742e6f938068"} Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.842801 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7bprj"] Nov 28 15:39:14 crc kubenswrapper[4805]: I1128 15:39:14.847442 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-7bprj"] Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.217940 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69b46aab-afdf-4551-b0b4-b8d4f8f97831" path="/var/lib/kubelet/pods/69b46aab-afdf-4551-b0b4-b8d4f8f97831/volumes" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.639427 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xcvf5"] Nov 28 15:39:15 crc kubenswrapper[4805]: E1128 15:39:15.640186 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b46aab-afdf-4551-b0b4-b8d4f8f97831" containerName="console" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.640218 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b46aab-afdf-4551-b0b4-b8d4f8f97831" containerName="console" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.640434 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="69b46aab-afdf-4551-b0b4-b8d4f8f97831" containerName="console" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.641766 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.665784 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xcvf5"] Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.797912 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-utilities\") pod \"redhat-operators-xcvf5\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.798651 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz4pv\" (UniqueName: \"kubernetes.io/projected/8c37e6c4-32d1-4cec-b914-d9a3db04b088-kube-api-access-vz4pv\") pod \"redhat-operators-xcvf5\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.798790 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-catalog-content\") pod \"redhat-operators-xcvf5\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.900087 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz4pv\" (UniqueName: \"kubernetes.io/projected/8c37e6c4-32d1-4cec-b914-d9a3db04b088-kube-api-access-vz4pv\") pod \"redhat-operators-xcvf5\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.900175 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-catalog-content\") pod \"redhat-operators-xcvf5\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.900247 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-utilities\") pod \"redhat-operators-xcvf5\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.900724 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-catalog-content\") pod \"redhat-operators-xcvf5\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.900817 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-utilities\") pod \"redhat-operators-xcvf5\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.928403 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz4pv\" (UniqueName: \"kubernetes.io/projected/8c37e6c4-32d1-4cec-b914-d9a3db04b088-kube-api-access-vz4pv\") pod \"redhat-operators-xcvf5\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:15 crc kubenswrapper[4805]: I1128 15:39:15.982210 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:16 crc kubenswrapper[4805]: I1128 15:39:16.208065 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xcvf5"] Nov 28 15:39:16 crc kubenswrapper[4805]: W1128 15:39:16.231592 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c37e6c4_32d1_4cec_b914_d9a3db04b088.slice/crio-2aa238d591dcaf6232711aea2207e8156dc2f608fa684333eb6587cc4f8ad162 WatchSource:0}: Error finding container 2aa238d591dcaf6232711aea2207e8156dc2f608fa684333eb6587cc4f8ad162: Status 404 returned error can't find the container with id 2aa238d591dcaf6232711aea2207e8156dc2f608fa684333eb6587cc4f8ad162 Nov 28 15:39:16 crc kubenswrapper[4805]: I1128 15:39:16.806842 4805 generic.go:334] "Generic (PLEG): container finished" podID="72ae900f-55af-457e-9785-8d671059454d" containerID="1d9d3924dac7f453321632c66723f2c3eec4dc032f7d77e7960725e5088acc0b" exitCode=0 Nov 28 15:39:16 crc kubenswrapper[4805]: I1128 15:39:16.806948 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" event={"ID":"72ae900f-55af-457e-9785-8d671059454d","Type":"ContainerDied","Data":"1d9d3924dac7f453321632c66723f2c3eec4dc032f7d77e7960725e5088acc0b"} Nov 28 15:39:16 crc kubenswrapper[4805]: I1128 15:39:16.810999 4805 generic.go:334] "Generic (PLEG): container finished" podID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerID="5e1c1646fc723104a580e2f8aa51d9854f87c13c2f02daf74404700559ca8aca" exitCode=0 Nov 28 15:39:16 crc kubenswrapper[4805]: I1128 15:39:16.811086 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcvf5" event={"ID":"8c37e6c4-32d1-4cec-b914-d9a3db04b088","Type":"ContainerDied","Data":"5e1c1646fc723104a580e2f8aa51d9854f87c13c2f02daf74404700559ca8aca"} Nov 28 15:39:16 crc kubenswrapper[4805]: I1128 15:39:16.811157 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcvf5" event={"ID":"8c37e6c4-32d1-4cec-b914-d9a3db04b088","Type":"ContainerStarted","Data":"2aa238d591dcaf6232711aea2207e8156dc2f608fa684333eb6587cc4f8ad162"} Nov 28 15:39:17 crc kubenswrapper[4805]: I1128 15:39:17.819511 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcvf5" event={"ID":"8c37e6c4-32d1-4cec-b914-d9a3db04b088","Type":"ContainerStarted","Data":"e32834c8eedaed1778c1a44d07a0ba056c06f0d157076042e7a6fae3d853ecc3"} Nov 28 15:39:17 crc kubenswrapper[4805]: I1128 15:39:17.823753 4805 generic.go:334] "Generic (PLEG): container finished" podID="72ae900f-55af-457e-9785-8d671059454d" containerID="4556ca06b70c07849c5e6c54c5728747db415085cbbfcdac06237247060c72f9" exitCode=0 Nov 28 15:39:17 crc kubenswrapper[4805]: I1128 15:39:17.823807 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" event={"ID":"72ae900f-55af-457e-9785-8d671059454d","Type":"ContainerDied","Data":"4556ca06b70c07849c5e6c54c5728747db415085cbbfcdac06237247060c72f9"} Nov 28 15:39:18 crc kubenswrapper[4805]: I1128 15:39:18.832646 4805 generic.go:334] "Generic (PLEG): container finished" podID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerID="e32834c8eedaed1778c1a44d07a0ba056c06f0d157076042e7a6fae3d853ecc3" exitCode=0 Nov 28 15:39:18 crc kubenswrapper[4805]: I1128 15:39:18.832690 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcvf5" event={"ID":"8c37e6c4-32d1-4cec-b914-d9a3db04b088","Type":"ContainerDied","Data":"e32834c8eedaed1778c1a44d07a0ba056c06f0d157076042e7a6fae3d853ecc3"} Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.142199 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.255857 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-util\") pod \"72ae900f-55af-457e-9785-8d671059454d\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.255983 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-bundle\") pod \"72ae900f-55af-457e-9785-8d671059454d\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.256041 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpr87\" (UniqueName: \"kubernetes.io/projected/72ae900f-55af-457e-9785-8d671059454d-kube-api-access-zpr87\") pod \"72ae900f-55af-457e-9785-8d671059454d\" (UID: \"72ae900f-55af-457e-9785-8d671059454d\") " Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.257104 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-bundle" (OuterVolumeSpecName: "bundle") pod "72ae900f-55af-457e-9785-8d671059454d" (UID: "72ae900f-55af-457e-9785-8d671059454d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.262582 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72ae900f-55af-457e-9785-8d671059454d-kube-api-access-zpr87" (OuterVolumeSpecName: "kube-api-access-zpr87") pod "72ae900f-55af-457e-9785-8d671059454d" (UID: "72ae900f-55af-457e-9785-8d671059454d"). InnerVolumeSpecName "kube-api-access-zpr87". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.274879 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-util" (OuterVolumeSpecName: "util") pod "72ae900f-55af-457e-9785-8d671059454d" (UID: "72ae900f-55af-457e-9785-8d671059454d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.357819 4805 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.357881 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpr87\" (UniqueName: \"kubernetes.io/projected/72ae900f-55af-457e-9785-8d671059454d-kube-api-access-zpr87\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.357905 4805 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ae900f-55af-457e-9785-8d671059454d-util\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.845151 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" event={"ID":"72ae900f-55af-457e-9785-8d671059454d","Type":"ContainerDied","Data":"759897d352b2b42ddf24a9a4ea5144867e8aa2ecb1266bf7bc46742e6f938068"} Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.845629 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="759897d352b2b42ddf24a9a4ea5144867e8aa2ecb1266bf7bc46742e6f938068" Nov 28 15:39:19 crc kubenswrapper[4805]: I1128 15:39:19.845214 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95" Nov 28 15:39:20 crc kubenswrapper[4805]: I1128 15:39:20.854793 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcvf5" event={"ID":"8c37e6c4-32d1-4cec-b914-d9a3db04b088","Type":"ContainerStarted","Data":"deeb6d15b284efb2ecbc8faddff7e465d41211201c5a6f1fdeaebbd93b8e184b"} Nov 28 15:39:20 crc kubenswrapper[4805]: I1128 15:39:20.899445 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xcvf5" podStartSLOduration=3.026195043 podStartE2EDuration="5.899413481s" podCreationTimestamp="2025-11-28 15:39:15 +0000 UTC" firstStartedPulling="2025-11-28 15:39:16.812783357 +0000 UTC m=+783.862574668" lastFinishedPulling="2025-11-28 15:39:19.686001765 +0000 UTC m=+786.735793106" observedRunningTime="2025-11-28 15:39:20.899047481 +0000 UTC m=+787.948838802" watchObservedRunningTime="2025-11-28 15:39:20.899413481 +0000 UTC m=+787.949204862" Nov 28 15:39:25 crc kubenswrapper[4805]: I1128 15:39:25.983238 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:25 crc kubenswrapper[4805]: I1128 15:39:25.983657 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:26 crc kubenswrapper[4805]: I1128 15:39:26.050115 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:26 crc kubenswrapper[4805]: I1128 15:39:26.939166 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:27 crc kubenswrapper[4805]: I1128 15:39:27.420840 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xcvf5"] Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.659496 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5"] Nov 28 15:39:28 crc kubenswrapper[4805]: E1128 15:39:28.659712 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ae900f-55af-457e-9785-8d671059454d" containerName="pull" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.659724 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ae900f-55af-457e-9785-8d671059454d" containerName="pull" Nov 28 15:39:28 crc kubenswrapper[4805]: E1128 15:39:28.659735 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ae900f-55af-457e-9785-8d671059454d" containerName="extract" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.659741 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ae900f-55af-457e-9785-8d671059454d" containerName="extract" Nov 28 15:39:28 crc kubenswrapper[4805]: E1128 15:39:28.659756 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ae900f-55af-457e-9785-8d671059454d" containerName="util" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.659762 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ae900f-55af-457e-9785-8d671059454d" containerName="util" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.659851 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ae900f-55af-457e-9785-8d671059454d" containerName="extract" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.660207 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.661876 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.662022 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-tjvq9" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.662209 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.662764 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.664135 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.678435 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5"] Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.800210 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/33947506-e060-4ec4-b7b1-dd52bdcbe7e7-apiservice-cert\") pod \"metallb-operator-controller-manager-8bcbb765d-vgkf5\" (UID: \"33947506-e060-4ec4-b7b1-dd52bdcbe7e7\") " pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.800267 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn8nb\" (UniqueName: \"kubernetes.io/projected/33947506-e060-4ec4-b7b1-dd52bdcbe7e7-kube-api-access-jn8nb\") pod \"metallb-operator-controller-manager-8bcbb765d-vgkf5\" (UID: \"33947506-e060-4ec4-b7b1-dd52bdcbe7e7\") " pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.800483 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/33947506-e060-4ec4-b7b1-dd52bdcbe7e7-webhook-cert\") pod \"metallb-operator-controller-manager-8bcbb765d-vgkf5\" (UID: \"33947506-e060-4ec4-b7b1-dd52bdcbe7e7\") " pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.901489 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/33947506-e060-4ec4-b7b1-dd52bdcbe7e7-apiservice-cert\") pod \"metallb-operator-controller-manager-8bcbb765d-vgkf5\" (UID: \"33947506-e060-4ec4-b7b1-dd52bdcbe7e7\") " pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.901544 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn8nb\" (UniqueName: \"kubernetes.io/projected/33947506-e060-4ec4-b7b1-dd52bdcbe7e7-kube-api-access-jn8nb\") pod \"metallb-operator-controller-manager-8bcbb765d-vgkf5\" (UID: \"33947506-e060-4ec4-b7b1-dd52bdcbe7e7\") " pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.901579 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/33947506-e060-4ec4-b7b1-dd52bdcbe7e7-webhook-cert\") pod \"metallb-operator-controller-manager-8bcbb765d-vgkf5\" (UID: \"33947506-e060-4ec4-b7b1-dd52bdcbe7e7\") " pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.903729 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xcvf5" podUID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerName="registry-server" containerID="cri-o://deeb6d15b284efb2ecbc8faddff7e465d41211201c5a6f1fdeaebbd93b8e184b" gracePeriod=2 Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.910488 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/33947506-e060-4ec4-b7b1-dd52bdcbe7e7-webhook-cert\") pod \"metallb-operator-controller-manager-8bcbb765d-vgkf5\" (UID: \"33947506-e060-4ec4-b7b1-dd52bdcbe7e7\") " pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.911028 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/33947506-e060-4ec4-b7b1-dd52bdcbe7e7-apiservice-cert\") pod \"metallb-operator-controller-manager-8bcbb765d-vgkf5\" (UID: \"33947506-e060-4ec4-b7b1-dd52bdcbe7e7\") " pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.919346 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn8nb\" (UniqueName: \"kubernetes.io/projected/33947506-e060-4ec4-b7b1-dd52bdcbe7e7-kube-api-access-jn8nb\") pod \"metallb-operator-controller-manager-8bcbb765d-vgkf5\" (UID: \"33947506-e060-4ec4-b7b1-dd52bdcbe7e7\") " pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:28 crc kubenswrapper[4805]: I1128 15:39:28.974851 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.104681 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb"] Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.106937 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.110169 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.111986 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.112640 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-rr4k4" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.116288 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb"] Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.204009 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/91105ea4-a3b0-41d6-be1e-73bf1d456a2c-apiservice-cert\") pod \"metallb-operator-webhook-server-5d55dc568c-ln8sb\" (UID: \"91105ea4-a3b0-41d6-be1e-73bf1d456a2c\") " pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.204562 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/91105ea4-a3b0-41d6-be1e-73bf1d456a2c-webhook-cert\") pod \"metallb-operator-webhook-server-5d55dc568c-ln8sb\" (UID: \"91105ea4-a3b0-41d6-be1e-73bf1d456a2c\") " pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.204624 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7zpw\" (UniqueName: \"kubernetes.io/projected/91105ea4-a3b0-41d6-be1e-73bf1d456a2c-kube-api-access-s7zpw\") pod \"metallb-operator-webhook-server-5d55dc568c-ln8sb\" (UID: \"91105ea4-a3b0-41d6-be1e-73bf1d456a2c\") " pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.238421 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5"] Nov 28 15:39:29 crc kubenswrapper[4805]: W1128 15:39:29.245596 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33947506_e060_4ec4_b7b1_dd52bdcbe7e7.slice/crio-e46c6ce9e9bba486d06cd50343557d9ee2c920393cbd654f2b8a92fc28faed82 WatchSource:0}: Error finding container e46c6ce9e9bba486d06cd50343557d9ee2c920393cbd654f2b8a92fc28faed82: Status 404 returned error can't find the container with id e46c6ce9e9bba486d06cd50343557d9ee2c920393cbd654f2b8a92fc28faed82 Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.306008 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/91105ea4-a3b0-41d6-be1e-73bf1d456a2c-webhook-cert\") pod \"metallb-operator-webhook-server-5d55dc568c-ln8sb\" (UID: \"91105ea4-a3b0-41d6-be1e-73bf1d456a2c\") " pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.306061 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7zpw\" (UniqueName: \"kubernetes.io/projected/91105ea4-a3b0-41d6-be1e-73bf1d456a2c-kube-api-access-s7zpw\") pod \"metallb-operator-webhook-server-5d55dc568c-ln8sb\" (UID: \"91105ea4-a3b0-41d6-be1e-73bf1d456a2c\") " pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.306104 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/91105ea4-a3b0-41d6-be1e-73bf1d456a2c-apiservice-cert\") pod \"metallb-operator-webhook-server-5d55dc568c-ln8sb\" (UID: \"91105ea4-a3b0-41d6-be1e-73bf1d456a2c\") " pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.310910 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/91105ea4-a3b0-41d6-be1e-73bf1d456a2c-apiservice-cert\") pod \"metallb-operator-webhook-server-5d55dc568c-ln8sb\" (UID: \"91105ea4-a3b0-41d6-be1e-73bf1d456a2c\") " pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.313967 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/91105ea4-a3b0-41d6-be1e-73bf1d456a2c-webhook-cert\") pod \"metallb-operator-webhook-server-5d55dc568c-ln8sb\" (UID: \"91105ea4-a3b0-41d6-be1e-73bf1d456a2c\") " pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.333354 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7zpw\" (UniqueName: \"kubernetes.io/projected/91105ea4-a3b0-41d6-be1e-73bf1d456a2c-kube-api-access-s7zpw\") pod \"metallb-operator-webhook-server-5d55dc568c-ln8sb\" (UID: \"91105ea4-a3b0-41d6-be1e-73bf1d456a2c\") " pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.450743 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.669190 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb"] Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.910507 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" event={"ID":"33947506-e060-4ec4-b7b1-dd52bdcbe7e7","Type":"ContainerStarted","Data":"e46c6ce9e9bba486d06cd50343557d9ee2c920393cbd654f2b8a92fc28faed82"} Nov 28 15:39:29 crc kubenswrapper[4805]: I1128 15:39:29.911485 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" event={"ID":"91105ea4-a3b0-41d6-be1e-73bf1d456a2c","Type":"ContainerStarted","Data":"5a923725f949c44cfc77d28f1fdbf13af1d5b8ed98a14e922fea18e2eaacff51"} Nov 28 15:39:31 crc kubenswrapper[4805]: I1128 15:39:31.924659 4805 generic.go:334] "Generic (PLEG): container finished" podID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerID="deeb6d15b284efb2ecbc8faddff7e465d41211201c5a6f1fdeaebbd93b8e184b" exitCode=0 Nov 28 15:39:31 crc kubenswrapper[4805]: I1128 15:39:31.924789 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcvf5" event={"ID":"8c37e6c4-32d1-4cec-b914-d9a3db04b088","Type":"ContainerDied","Data":"deeb6d15b284efb2ecbc8faddff7e465d41211201c5a6f1fdeaebbd93b8e184b"} Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.763234 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.857649 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz4pv\" (UniqueName: \"kubernetes.io/projected/8c37e6c4-32d1-4cec-b914-d9a3db04b088-kube-api-access-vz4pv\") pod \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.857893 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-catalog-content\") pod \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.857971 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-utilities\") pod \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\" (UID: \"8c37e6c4-32d1-4cec-b914-d9a3db04b088\") " Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.858931 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-utilities" (OuterVolumeSpecName: "utilities") pod "8c37e6c4-32d1-4cec-b914-d9a3db04b088" (UID: "8c37e6c4-32d1-4cec-b914-d9a3db04b088"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.872543 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c37e6c4-32d1-4cec-b914-d9a3db04b088-kube-api-access-vz4pv" (OuterVolumeSpecName: "kube-api-access-vz4pv") pod "8c37e6c4-32d1-4cec-b914-d9a3db04b088" (UID: "8c37e6c4-32d1-4cec-b914-d9a3db04b088"). InnerVolumeSpecName "kube-api-access-vz4pv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.947910 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcvf5" event={"ID":"8c37e6c4-32d1-4cec-b914-d9a3db04b088","Type":"ContainerDied","Data":"2aa238d591dcaf6232711aea2207e8156dc2f608fa684333eb6587cc4f8ad162"} Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.947962 4805 scope.go:117] "RemoveContainer" containerID="deeb6d15b284efb2ecbc8faddff7e465d41211201c5a6f1fdeaebbd93b8e184b" Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.948082 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcvf5" Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.960506 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.960531 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz4pv\" (UniqueName: \"kubernetes.io/projected/8c37e6c4-32d1-4cec-b914-d9a3db04b088-kube-api-access-vz4pv\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.964178 4805 scope.go:117] "RemoveContainer" containerID="e32834c8eedaed1778c1a44d07a0ba056c06f0d157076042e7a6fae3d853ecc3" Nov 28 15:39:32 crc kubenswrapper[4805]: I1128 15:39:32.984511 4805 scope.go:117] "RemoveContainer" containerID="5e1c1646fc723104a580e2f8aa51d9854f87c13c2f02daf74404700559ca8aca" Nov 28 15:39:33 crc kubenswrapper[4805]: I1128 15:39:33.008898 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c37e6c4-32d1-4cec-b914-d9a3db04b088" (UID: "8c37e6c4-32d1-4cec-b914-d9a3db04b088"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:39:33 crc kubenswrapper[4805]: I1128 15:39:33.061719 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c37e6c4-32d1-4cec-b914-d9a3db04b088-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:33 crc kubenswrapper[4805]: I1128 15:39:33.287495 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xcvf5"] Nov 28 15:39:33 crc kubenswrapper[4805]: I1128 15:39:33.291157 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xcvf5"] Nov 28 15:39:33 crc kubenswrapper[4805]: I1128 15:39:33.954669 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" event={"ID":"33947506-e060-4ec4-b7b1-dd52bdcbe7e7","Type":"ContainerStarted","Data":"a7817ca4b40d2d8494e875216c3bfd008528d5347b320a426edc0d0b616cb5ea"} Nov 28 15:39:33 crc kubenswrapper[4805]: I1128 15:39:33.955512 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:39:33 crc kubenswrapper[4805]: I1128 15:39:33.973116 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" podStartSLOduration=2.447161554 podStartE2EDuration="5.97309816s" podCreationTimestamp="2025-11-28 15:39:28 +0000 UTC" firstStartedPulling="2025-11-28 15:39:29.258625471 +0000 UTC m=+796.308416792" lastFinishedPulling="2025-11-28 15:39:32.784562087 +0000 UTC m=+799.834353398" observedRunningTime="2025-11-28 15:39:33.97049211 +0000 UTC m=+801.020283431" watchObservedRunningTime="2025-11-28 15:39:33.97309816 +0000 UTC m=+801.022889471" Nov 28 15:39:35 crc kubenswrapper[4805]: I1128 15:39:35.211449 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" path="/var/lib/kubelet/pods/8c37e6c4-32d1-4cec-b914-d9a3db04b088/volumes" Nov 28 15:39:35 crc kubenswrapper[4805]: I1128 15:39:35.969154 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" event={"ID":"91105ea4-a3b0-41d6-be1e-73bf1d456a2c","Type":"ContainerStarted","Data":"4028892061fe5b077d7bce513fff65ae622cfad04aeb76020d6c9e5eb96a9ebd"} Nov 28 15:39:35 crc kubenswrapper[4805]: I1128 15:39:35.969582 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:39:35 crc kubenswrapper[4805]: I1128 15:39:35.993826 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" podStartSLOduration=1.4695131 podStartE2EDuration="6.99379826s" podCreationTimestamp="2025-11-28 15:39:29 +0000 UTC" firstStartedPulling="2025-11-28 15:39:29.674279573 +0000 UTC m=+796.724070884" lastFinishedPulling="2025-11-28 15:39:35.198564723 +0000 UTC m=+802.248356044" observedRunningTime="2025-11-28 15:39:35.987938731 +0000 UTC m=+803.037730062" watchObservedRunningTime="2025-11-28 15:39:35.99379826 +0000 UTC m=+803.043589611" Nov 28 15:39:49 crc kubenswrapper[4805]: I1128 15:39:49.458022 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5d55dc568c-ln8sb" Nov 28 15:40:08 crc kubenswrapper[4805]: I1128 15:40:08.977293 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-8bcbb765d-vgkf5" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.678257 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85"] Nov 28 15:40:09 crc kubenswrapper[4805]: E1128 15:40:09.678550 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerName="extract-utilities" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.678575 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerName="extract-utilities" Nov 28 15:40:09 crc kubenswrapper[4805]: E1128 15:40:09.678588 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerName="extract-content" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.678596 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerName="extract-content" Nov 28 15:40:09 crc kubenswrapper[4805]: E1128 15:40:09.678622 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerName="registry-server" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.678630 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerName="registry-server" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.678750 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c37e6c4-32d1-4cec-b914-d9a3db04b088" containerName="registry-server" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.679262 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.682753 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.682904 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-7fphp" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.682977 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-l9qgw"] Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.685734 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.687464 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.687974 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.690293 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85"] Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.770760 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6wzdr"] Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.771584 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.775747 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-mlmn8" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.775816 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.776466 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.778712 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.793796 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-lkhkl"] Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.794640 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.799498 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.808242 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-lkhkl"] Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.846797 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8aa6b614-d8e7-4956-afc5-3b30a530b76f-frr-startup\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.846871 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-memberlist\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.846893 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7tsc\" (UniqueName: \"kubernetes.io/projected/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-kube-api-access-b7tsc\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.846912 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrzrb\" (UniqueName: \"kubernetes.io/projected/8aa6b614-d8e7-4956-afc5-3b30a530b76f-kube-api-access-xrzrb\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.846931 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0f9d1f71-55cc-4bc7-827f-a1b2c12da36a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-znh85\" (UID: \"0f9d1f71-55cc-4bc7-827f-a1b2c12da36a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.846977 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-frr-sockets\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.846997 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x848h\" (UniqueName: \"kubernetes.io/projected/0f9d1f71-55cc-4bc7-827f-a1b2c12da36a-kube-api-access-x848h\") pod \"frr-k8s-webhook-server-7fcb986d4-znh85\" (UID: \"0f9d1f71-55cc-4bc7-827f-a1b2c12da36a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.847020 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-reloader\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.847040 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-metallb-excludel2\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.847060 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-frr-conf\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.847204 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-metrics-certs\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.847265 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8aa6b614-d8e7-4956-afc5-3b30a530b76f-metrics-certs\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.847345 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-metrics\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948579 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-frr-conf\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948666 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr4zn\" (UniqueName: \"kubernetes.io/projected/4265aa27-789b-4c6c-9d99-e061e507aea2-kube-api-access-tr4zn\") pod \"controller-f8648f98b-lkhkl\" (UID: \"4265aa27-789b-4c6c-9d99-e061e507aea2\") " pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948694 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-metrics-certs\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948726 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8aa6b614-d8e7-4956-afc5-3b30a530b76f-metrics-certs\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948752 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-metrics\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948795 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8aa6b614-d8e7-4956-afc5-3b30a530b76f-frr-startup\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948828 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-memberlist\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948854 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7tsc\" (UniqueName: \"kubernetes.io/projected/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-kube-api-access-b7tsc\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948878 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrzrb\" (UniqueName: \"kubernetes.io/projected/8aa6b614-d8e7-4956-afc5-3b30a530b76f-kube-api-access-xrzrb\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948904 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0f9d1f71-55cc-4bc7-827f-a1b2c12da36a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-znh85\" (UID: \"0f9d1f71-55cc-4bc7-827f-a1b2c12da36a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948930 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-frr-sockets\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948953 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x848h\" (UniqueName: \"kubernetes.io/projected/0f9d1f71-55cc-4bc7-827f-a1b2c12da36a-kube-api-access-x848h\") pod \"frr-k8s-webhook-server-7fcb986d4-znh85\" (UID: \"0f9d1f71-55cc-4bc7-827f-a1b2c12da36a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.948979 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4265aa27-789b-4c6c-9d99-e061e507aea2-metrics-certs\") pod \"controller-f8648f98b-lkhkl\" (UID: \"4265aa27-789b-4c6c-9d99-e061e507aea2\") " pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.949009 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-reloader\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.949032 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-metallb-excludel2\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.949049 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4265aa27-789b-4c6c-9d99-e061e507aea2-cert\") pod \"controller-f8648f98b-lkhkl\" (UID: \"4265aa27-789b-4c6c-9d99-e061e507aea2\") " pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.949790 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-frr-conf\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.949984 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-reloader\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.950256 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-frr-sockets\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.950481 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-metallb-excludel2\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: E1128 15:40:09.950568 4805 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 15:40:09 crc kubenswrapper[4805]: E1128 15:40:09.950635 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-memberlist podName:85167fd2-c5cc-4c8e-9106-5912aa9d4dee nodeName:}" failed. No retries permitted until 2025-11-28 15:40:10.450616908 +0000 UTC m=+837.500408309 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-memberlist") pod "speaker-6wzdr" (UID: "85167fd2-c5cc-4c8e-9106-5912aa9d4dee") : secret "metallb-memberlist" not found Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.950680 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8aa6b614-d8e7-4956-afc5-3b30a530b76f-metrics\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.951157 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8aa6b614-d8e7-4956-afc5-3b30a530b76f-frr-startup\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.959002 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-metrics-certs\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.959026 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0f9d1f71-55cc-4bc7-827f-a1b2c12da36a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-znh85\" (UID: \"0f9d1f71-55cc-4bc7-827f-a1b2c12da36a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.959741 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8aa6b614-d8e7-4956-afc5-3b30a530b76f-metrics-certs\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.972324 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7tsc\" (UniqueName: \"kubernetes.io/projected/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-kube-api-access-b7tsc\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.976670 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrzrb\" (UniqueName: \"kubernetes.io/projected/8aa6b614-d8e7-4956-afc5-3b30a530b76f-kube-api-access-xrzrb\") pod \"frr-k8s-l9qgw\" (UID: \"8aa6b614-d8e7-4956-afc5-3b30a530b76f\") " pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.979992 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x848h\" (UniqueName: \"kubernetes.io/projected/0f9d1f71-55cc-4bc7-827f-a1b2c12da36a-kube-api-access-x848h\") pod \"frr-k8s-webhook-server-7fcb986d4-znh85\" (UID: \"0f9d1f71-55cc-4bc7-827f-a1b2c12da36a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:09 crc kubenswrapper[4805]: I1128 15:40:09.995723 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.004072 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.050220 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4265aa27-789b-4c6c-9d99-e061e507aea2-cert\") pod \"controller-f8648f98b-lkhkl\" (UID: \"4265aa27-789b-4c6c-9d99-e061e507aea2\") " pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.050299 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr4zn\" (UniqueName: \"kubernetes.io/projected/4265aa27-789b-4c6c-9d99-e061e507aea2-kube-api-access-tr4zn\") pod \"controller-f8648f98b-lkhkl\" (UID: \"4265aa27-789b-4c6c-9d99-e061e507aea2\") " pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.050405 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4265aa27-789b-4c6c-9d99-e061e507aea2-metrics-certs\") pod \"controller-f8648f98b-lkhkl\" (UID: \"4265aa27-789b-4c6c-9d99-e061e507aea2\") " pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.054471 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4265aa27-789b-4c6c-9d99-e061e507aea2-cert\") pod \"controller-f8648f98b-lkhkl\" (UID: \"4265aa27-789b-4c6c-9d99-e061e507aea2\") " pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.056794 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4265aa27-789b-4c6c-9d99-e061e507aea2-metrics-certs\") pod \"controller-f8648f98b-lkhkl\" (UID: \"4265aa27-789b-4c6c-9d99-e061e507aea2\") " pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.068159 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr4zn\" (UniqueName: \"kubernetes.io/projected/4265aa27-789b-4c6c-9d99-e061e507aea2-kube-api-access-tr4zn\") pod \"controller-f8648f98b-lkhkl\" (UID: \"4265aa27-789b-4c6c-9d99-e061e507aea2\") " pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.109815 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:10 crc kubenswrapper[4805]: W1128 15:40:10.239143 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f9d1f71_55cc_4bc7_827f_a1b2c12da36a.slice/crio-d879b6263037d09de3a6a00dbe6f17133d0f0e3f49dbc28e93bf4686279324d0 WatchSource:0}: Error finding container d879b6263037d09de3a6a00dbe6f17133d0f0e3f49dbc28e93bf4686279324d0: Status 404 returned error can't find the container with id d879b6263037d09de3a6a00dbe6f17133d0f0e3f49dbc28e93bf4686279324d0 Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.239952 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85"] Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.311512 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-lkhkl"] Nov 28 15:40:10 crc kubenswrapper[4805]: W1128 15:40:10.315495 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4265aa27_789b_4c6c_9d99_e061e507aea2.slice/crio-ff69756c51bbe4e17fd4d86060e9da3e6eaa6c8bf72845865918734d8d061aac WatchSource:0}: Error finding container ff69756c51bbe4e17fd4d86060e9da3e6eaa6c8bf72845865918734d8d061aac: Status 404 returned error can't find the container with id ff69756c51bbe4e17fd4d86060e9da3e6eaa6c8bf72845865918734d8d061aac Nov 28 15:40:10 crc kubenswrapper[4805]: I1128 15:40:10.463573 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-memberlist\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:10 crc kubenswrapper[4805]: E1128 15:40:10.463972 4805 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 15:40:10 crc kubenswrapper[4805]: E1128 15:40:10.464059 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-memberlist podName:85167fd2-c5cc-4c8e-9106-5912aa9d4dee nodeName:}" failed. No retries permitted until 2025-11-28 15:40:11.464037586 +0000 UTC m=+838.513828937 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-memberlist") pod "speaker-6wzdr" (UID: "85167fd2-c5cc-4c8e-9106-5912aa9d4dee") : secret "metallb-memberlist" not found Nov 28 15:40:11 crc kubenswrapper[4805]: I1128 15:40:11.194191 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerStarted","Data":"98a5d22e9b9c94170c56a976f25d86a30d61f1606707ba10408b880369a985b3"} Nov 28 15:40:11 crc kubenswrapper[4805]: I1128 15:40:11.195718 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" event={"ID":"0f9d1f71-55cc-4bc7-827f-a1b2c12da36a","Type":"ContainerStarted","Data":"d879b6263037d09de3a6a00dbe6f17133d0f0e3f49dbc28e93bf4686279324d0"} Nov 28 15:40:11 crc kubenswrapper[4805]: I1128 15:40:11.198023 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-lkhkl" event={"ID":"4265aa27-789b-4c6c-9d99-e061e507aea2","Type":"ContainerStarted","Data":"ff69756c51bbe4e17fd4d86060e9da3e6eaa6c8bf72845865918734d8d061aac"} Nov 28 15:40:11 crc kubenswrapper[4805]: I1128 15:40:11.479584 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-memberlist\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:11 crc kubenswrapper[4805]: I1128 15:40:11.487524 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/85167fd2-c5cc-4c8e-9106-5912aa9d4dee-memberlist\") pod \"speaker-6wzdr\" (UID: \"85167fd2-c5cc-4c8e-9106-5912aa9d4dee\") " pod="metallb-system/speaker-6wzdr" Nov 28 15:40:11 crc kubenswrapper[4805]: I1128 15:40:11.582944 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6wzdr" Nov 28 15:40:12 crc kubenswrapper[4805]: I1128 15:40:12.207965 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-lkhkl" event={"ID":"4265aa27-789b-4c6c-9d99-e061e507aea2","Type":"ContainerStarted","Data":"7ca5d9fd6e75a73eebf978168a7fc1430f983c69f2402dc50ffb6bef10e30f28"} Nov 28 15:40:12 crc kubenswrapper[4805]: I1128 15:40:12.208334 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:12 crc kubenswrapper[4805]: I1128 15:40:12.208350 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-lkhkl" event={"ID":"4265aa27-789b-4c6c-9d99-e061e507aea2","Type":"ContainerStarted","Data":"65adcaeff3b73835ed2a5b0469cb036c9c7ff5dd1e83f20dc808545016c26f1f"} Nov 28 15:40:12 crc kubenswrapper[4805]: I1128 15:40:12.210442 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6wzdr" event={"ID":"85167fd2-c5cc-4c8e-9106-5912aa9d4dee","Type":"ContainerStarted","Data":"74d8e7fc52db7e4b56de686d607c84c08bd2098dbc032692594a9f03bdc9320a"} Nov 28 15:40:12 crc kubenswrapper[4805]: I1128 15:40:12.210476 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6wzdr" event={"ID":"85167fd2-c5cc-4c8e-9106-5912aa9d4dee","Type":"ContainerStarted","Data":"908abfb288aac21ca320f0a4b9fc96d6167a96132a6484720f9622f0bfbda437"} Nov 28 15:40:12 crc kubenswrapper[4805]: I1128 15:40:12.210485 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6wzdr" event={"ID":"85167fd2-c5cc-4c8e-9106-5912aa9d4dee","Type":"ContainerStarted","Data":"5ba1c6e0865ba4490882788e2474e39d454c2e0c664ba94845d5f2a86aed65da"} Nov 28 15:40:12 crc kubenswrapper[4805]: I1128 15:40:12.210800 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6wzdr" Nov 28 15:40:12 crc kubenswrapper[4805]: I1128 15:40:12.229344 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-lkhkl" podStartSLOduration=3.229324991 podStartE2EDuration="3.229324991s" podCreationTimestamp="2025-11-28 15:40:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:40:12.228694024 +0000 UTC m=+839.278485345" watchObservedRunningTime="2025-11-28 15:40:12.229324991 +0000 UTC m=+839.279116292" Nov 28 15:40:12 crc kubenswrapper[4805]: I1128 15:40:12.245437 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6wzdr" podStartSLOduration=3.245416117 podStartE2EDuration="3.245416117s" podCreationTimestamp="2025-11-28 15:40:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:40:12.244599185 +0000 UTC m=+839.294390486" watchObservedRunningTime="2025-11-28 15:40:12.245416117 +0000 UTC m=+839.295207428" Nov 28 15:40:18 crc kubenswrapper[4805]: I1128 15:40:18.255074 4805 generic.go:334] "Generic (PLEG): container finished" podID="8aa6b614-d8e7-4956-afc5-3b30a530b76f" containerID="aa59a7c7ae996d8dd59e9c5ec4e2e929021c88414a95231a41db2b983054041a" exitCode=0 Nov 28 15:40:18 crc kubenswrapper[4805]: I1128 15:40:18.255183 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerDied","Data":"aa59a7c7ae996d8dd59e9c5ec4e2e929021c88414a95231a41db2b983054041a"} Nov 28 15:40:18 crc kubenswrapper[4805]: I1128 15:40:18.258234 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" event={"ID":"0f9d1f71-55cc-4bc7-827f-a1b2c12da36a","Type":"ContainerStarted","Data":"78346532fd0fa2244b028c919857a9748c36e2a8313cca497846f542580d2ed7"} Nov 28 15:40:18 crc kubenswrapper[4805]: I1128 15:40:18.258533 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:18 crc kubenswrapper[4805]: I1128 15:40:18.318044 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" podStartSLOduration=2.473008151 podStartE2EDuration="9.318018582s" podCreationTimestamp="2025-11-28 15:40:09 +0000 UTC" firstStartedPulling="2025-11-28 15:40:10.243063504 +0000 UTC m=+837.292854815" lastFinishedPulling="2025-11-28 15:40:17.088073935 +0000 UTC m=+844.137865246" observedRunningTime="2025-11-28 15:40:18.312793214 +0000 UTC m=+845.362584545" watchObservedRunningTime="2025-11-28 15:40:18.318018582 +0000 UTC m=+845.367809913" Nov 28 15:40:19 crc kubenswrapper[4805]: I1128 15:40:19.266524 4805 generic.go:334] "Generic (PLEG): container finished" podID="8aa6b614-d8e7-4956-afc5-3b30a530b76f" containerID="7703275d9fcd943c1fea976c9cb6631eb71467ea286f3098e48a436053d73ccb" exitCode=0 Nov 28 15:40:19 crc kubenswrapper[4805]: I1128 15:40:19.266636 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerDied","Data":"7703275d9fcd943c1fea976c9cb6631eb71467ea286f3098e48a436053d73ccb"} Nov 28 15:40:20 crc kubenswrapper[4805]: I1128 15:40:20.277792 4805 generic.go:334] "Generic (PLEG): container finished" podID="8aa6b614-d8e7-4956-afc5-3b30a530b76f" containerID="ebc2c8940476d65a4b50ddfffa80f3df697b8357477d389aad468d0420b61564" exitCode=0 Nov 28 15:40:20 crc kubenswrapper[4805]: I1128 15:40:20.277857 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerDied","Data":"ebc2c8940476d65a4b50ddfffa80f3df697b8357477d389aad468d0420b61564"} Nov 28 15:40:21 crc kubenswrapper[4805]: I1128 15:40:21.286129 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerStarted","Data":"d310d41d8d43382f64926afdd89c3bd2a74e182d016f8e6a306db59a0961836d"} Nov 28 15:40:21 crc kubenswrapper[4805]: I1128 15:40:21.286482 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerStarted","Data":"a1f22b166afcea28e9fd1f4948ec4ce63ec0353bdaee5b00910c34a3ec3bb107"} Nov 28 15:40:21 crc kubenswrapper[4805]: I1128 15:40:21.286494 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerStarted","Data":"a966cb93bb3f1abefe519c14419c346bdbadb7b8ddab9cee827680248f400dc4"} Nov 28 15:40:21 crc kubenswrapper[4805]: I1128 15:40:21.286502 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerStarted","Data":"1f1d416b575801184a8bee467aba346be922d5ad8e454af8cd0263c786c66677"} Nov 28 15:40:21 crc kubenswrapper[4805]: I1128 15:40:21.286512 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerStarted","Data":"3e04f37432fcb33b669421644acd92b437620f67806dc96398b2ba15657f2560"} Nov 28 15:40:21 crc kubenswrapper[4805]: I1128 15:40:21.589706 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6wzdr" Nov 28 15:40:22 crc kubenswrapper[4805]: I1128 15:40:22.294558 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l9qgw" event={"ID":"8aa6b614-d8e7-4956-afc5-3b30a530b76f","Type":"ContainerStarted","Data":"f227c74ad6569dd1d8fa9624cbb65270d6a46656ca9c8fd95c7933c6fb8f4801"} Nov 28 15:40:22 crc kubenswrapper[4805]: I1128 15:40:22.294787 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:22 crc kubenswrapper[4805]: I1128 15:40:22.318119 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-l9qgw" podStartSLOduration=7.380423511 podStartE2EDuration="13.318098241s" podCreationTimestamp="2025-11-28 15:40:09 +0000 UTC" firstStartedPulling="2025-11-28 15:40:11.183993384 +0000 UTC m=+838.233784705" lastFinishedPulling="2025-11-28 15:40:17.121668124 +0000 UTC m=+844.171459435" observedRunningTime="2025-11-28 15:40:22.314577688 +0000 UTC m=+849.364369019" watchObservedRunningTime="2025-11-28 15:40:22.318098241 +0000 UTC m=+849.367889552" Nov 28 15:40:22 crc kubenswrapper[4805]: I1128 15:40:22.995804 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz"] Nov 28 15:40:22 crc kubenswrapper[4805]: I1128 15:40:22.997181 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:22 crc kubenswrapper[4805]: I1128 15:40:22.999162 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.010208 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz"] Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.077063 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.077121 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.077176 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znhww\" (UniqueName: \"kubernetes.io/projected/4c270e31-7a0d-4985-b559-6d4ebc5440a2-kube-api-access-znhww\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.178198 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.178268 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.178321 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znhww\" (UniqueName: \"kubernetes.io/projected/4c270e31-7a0d-4985-b559-6d4ebc5440a2-kube-api-access-znhww\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.178828 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.178839 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.203256 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znhww\" (UniqueName: \"kubernetes.io/projected/4c270e31-7a0d-4985-b559-6d4ebc5440a2-kube-api-access-znhww\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.316729 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:23 crc kubenswrapper[4805]: I1128 15:40:23.728249 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz"] Nov 28 15:40:24 crc kubenswrapper[4805]: I1128 15:40:24.306983 4805 generic.go:334] "Generic (PLEG): container finished" podID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerID="b9b499e17be43713f22b7fb28351b325ce0b918f140e7495278e6916a6e1d5fa" exitCode=0 Nov 28 15:40:24 crc kubenswrapper[4805]: I1128 15:40:24.307033 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" event={"ID":"4c270e31-7a0d-4985-b559-6d4ebc5440a2","Type":"ContainerDied","Data":"b9b499e17be43713f22b7fb28351b325ce0b918f140e7495278e6916a6e1d5fa"} Nov 28 15:40:24 crc kubenswrapper[4805]: I1128 15:40:24.307316 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" event={"ID":"4c270e31-7a0d-4985-b559-6d4ebc5440a2","Type":"ContainerStarted","Data":"4610a5bac01db89a18b19aa7bd7a81543656f308dab3ad6cd1aec066f65f4c23"} Nov 28 15:40:25 crc kubenswrapper[4805]: I1128 15:40:25.005529 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:25 crc kubenswrapper[4805]: I1128 15:40:25.066044 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:30 crc kubenswrapper[4805]: I1128 15:40:30.004228 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-znh85" Nov 28 15:40:30 crc kubenswrapper[4805]: I1128 15:40:30.007463 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-l9qgw" Nov 28 15:40:30 crc kubenswrapper[4805]: I1128 15:40:30.115692 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-lkhkl" Nov 28 15:40:30 crc kubenswrapper[4805]: I1128 15:40:30.347143 4805 generic.go:334] "Generic (PLEG): container finished" podID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerID="00584f4ed1577bc5a4b62e87e0bbfe5f0294c3c8323d6e631f268d0533cc9d9c" exitCode=0 Nov 28 15:40:30 crc kubenswrapper[4805]: I1128 15:40:30.347180 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" event={"ID":"4c270e31-7a0d-4985-b559-6d4ebc5440a2","Type":"ContainerDied","Data":"00584f4ed1577bc5a4b62e87e0bbfe5f0294c3c8323d6e631f268d0533cc9d9c"} Nov 28 15:40:31 crc kubenswrapper[4805]: I1128 15:40:31.357934 4805 generic.go:334] "Generic (PLEG): container finished" podID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerID="c95dbfc81f9adaf43f3d433afd9733628e5c2208d1ce5bbdfe460b33cd517526" exitCode=0 Nov 28 15:40:31 crc kubenswrapper[4805]: I1128 15:40:31.358263 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" event={"ID":"4c270e31-7a0d-4985-b559-6d4ebc5440a2","Type":"ContainerDied","Data":"c95dbfc81f9adaf43f3d433afd9733628e5c2208d1ce5bbdfe460b33cd517526"} Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.635787 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.763827 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znhww\" (UniqueName: \"kubernetes.io/projected/4c270e31-7a0d-4985-b559-6d4ebc5440a2-kube-api-access-znhww\") pod \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.763927 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-bundle\") pod \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.763963 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-util\") pod \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\" (UID: \"4c270e31-7a0d-4985-b559-6d4ebc5440a2\") " Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.765677 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-bundle" (OuterVolumeSpecName: "bundle") pod "4c270e31-7a0d-4985-b559-6d4ebc5440a2" (UID: "4c270e31-7a0d-4985-b559-6d4ebc5440a2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.771693 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c270e31-7a0d-4985-b559-6d4ebc5440a2-kube-api-access-znhww" (OuterVolumeSpecName: "kube-api-access-znhww") pod "4c270e31-7a0d-4985-b559-6d4ebc5440a2" (UID: "4c270e31-7a0d-4985-b559-6d4ebc5440a2"). InnerVolumeSpecName "kube-api-access-znhww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.786549 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-util" (OuterVolumeSpecName: "util") pod "4c270e31-7a0d-4985-b559-6d4ebc5440a2" (UID: "4c270e31-7a0d-4985-b559-6d4ebc5440a2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.866015 4805 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-util\") on node \"crc\" DevicePath \"\"" Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.866065 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znhww\" (UniqueName: \"kubernetes.io/projected/4c270e31-7a0d-4985-b559-6d4ebc5440a2-kube-api-access-znhww\") on node \"crc\" DevicePath \"\"" Nov 28 15:40:32 crc kubenswrapper[4805]: I1128 15:40:32.866080 4805 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c270e31-7a0d-4985-b559-6d4ebc5440a2-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:40:33 crc kubenswrapper[4805]: I1128 15:40:33.373917 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" event={"ID":"4c270e31-7a0d-4985-b559-6d4ebc5440a2","Type":"ContainerDied","Data":"4610a5bac01db89a18b19aa7bd7a81543656f308dab3ad6cd1aec066f65f4c23"} Nov 28 15:40:33 crc kubenswrapper[4805]: I1128 15:40:33.373965 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz" Nov 28 15:40:33 crc kubenswrapper[4805]: I1128 15:40:33.373971 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4610a5bac01db89a18b19aa7bd7a81543656f308dab3ad6cd1aec066f65f4c23" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.467990 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b"] Nov 28 15:40:41 crc kubenswrapper[4805]: E1128 15:40:41.468891 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerName="util" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.468907 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerName="util" Nov 28 15:40:41 crc kubenswrapper[4805]: E1128 15:40:41.468921 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerName="pull" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.468928 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerName="pull" Nov 28 15:40:41 crc kubenswrapper[4805]: E1128 15:40:41.468942 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerName="extract" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.468950 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerName="extract" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.469083 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c270e31-7a0d-4985-b559-6d4ebc5440a2" containerName="extract" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.469596 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.471142 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.471487 4805 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-hfgfb" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.471622 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.520918 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b"] Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.624988 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbs8v\" (UniqueName: \"kubernetes.io/projected/762fbb57-9692-4671-b733-694d30caac6c-kube-api-access-jbs8v\") pod \"cert-manager-operator-controller-manager-64cf6dff88-mfm6b\" (UID: \"762fbb57-9692-4671-b733-694d30caac6c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.625048 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/762fbb57-9692-4671-b733-694d30caac6c-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-mfm6b\" (UID: \"762fbb57-9692-4671-b733-694d30caac6c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.726082 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbs8v\" (UniqueName: \"kubernetes.io/projected/762fbb57-9692-4671-b733-694d30caac6c-kube-api-access-jbs8v\") pod \"cert-manager-operator-controller-manager-64cf6dff88-mfm6b\" (UID: \"762fbb57-9692-4671-b733-694d30caac6c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.726153 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/762fbb57-9692-4671-b733-694d30caac6c-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-mfm6b\" (UID: \"762fbb57-9692-4671-b733-694d30caac6c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.726771 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/762fbb57-9692-4671-b733-694d30caac6c-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-mfm6b\" (UID: \"762fbb57-9692-4671-b733-694d30caac6c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.751179 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbs8v\" (UniqueName: \"kubernetes.io/projected/762fbb57-9692-4671-b733-694d30caac6c-kube-api-access-jbs8v\") pod \"cert-manager-operator-controller-manager-64cf6dff88-mfm6b\" (UID: \"762fbb57-9692-4671-b733-694d30caac6c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" Nov 28 15:40:41 crc kubenswrapper[4805]: I1128 15:40:41.785464 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" Nov 28 15:40:42 crc kubenswrapper[4805]: I1128 15:40:42.239600 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b"] Nov 28 15:40:42 crc kubenswrapper[4805]: I1128 15:40:42.426651 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" event={"ID":"762fbb57-9692-4671-b733-694d30caac6c","Type":"ContainerStarted","Data":"1c9aa2f754f207230df8edd2e50f3afff5078350f77ab30b5033f71662b9984d"} Nov 28 15:40:50 crc kubenswrapper[4805]: I1128 15:40:50.479928 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" event={"ID":"762fbb57-9692-4671-b733-694d30caac6c","Type":"ContainerStarted","Data":"d2391c0600c210e5d480a9a6c0f91b9e2a5fe03880977b26b8864036cd172baf"} Nov 28 15:40:50 crc kubenswrapper[4805]: I1128 15:40:50.506185 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-mfm6b" podStartSLOduration=2.095986166 podStartE2EDuration="9.506154884s" podCreationTimestamp="2025-11-28 15:40:41 +0000 UTC" firstStartedPulling="2025-11-28 15:40:42.26335205 +0000 UTC m=+869.313143371" lastFinishedPulling="2025-11-28 15:40:49.673520768 +0000 UTC m=+876.723312089" observedRunningTime="2025-11-28 15:40:50.501591203 +0000 UTC m=+877.551382584" watchObservedRunningTime="2025-11-28 15:40:50.506154884 +0000 UTC m=+877.555946235" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.639986 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-rqb4d"] Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.641301 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.644265 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.644597 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.646931 4805 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-tncgb" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.651526 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-rqb4d"] Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.724341 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jd8z\" (UniqueName: \"kubernetes.io/projected/8a45f875-20b4-409a-b77a-22305360aa93-kube-api-access-8jd8z\") pod \"cert-manager-webhook-f4fb5df64-rqb4d\" (UID: \"8a45f875-20b4-409a-b77a-22305360aa93\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.724439 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a45f875-20b4-409a-b77a-22305360aa93-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-rqb4d\" (UID: \"8a45f875-20b4-409a-b77a-22305360aa93\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.825238 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jd8z\" (UniqueName: \"kubernetes.io/projected/8a45f875-20b4-409a-b77a-22305360aa93-kube-api-access-8jd8z\") pod \"cert-manager-webhook-f4fb5df64-rqb4d\" (UID: \"8a45f875-20b4-409a-b77a-22305360aa93\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.825312 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a45f875-20b4-409a-b77a-22305360aa93-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-rqb4d\" (UID: \"8a45f875-20b4-409a-b77a-22305360aa93\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.844393 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a45f875-20b4-409a-b77a-22305360aa93-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-rqb4d\" (UID: \"8a45f875-20b4-409a-b77a-22305360aa93\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.846190 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jd8z\" (UniqueName: \"kubernetes.io/projected/8a45f875-20b4-409a-b77a-22305360aa93-kube-api-access-8jd8z\") pod \"cert-manager-webhook-f4fb5df64-rqb4d\" (UID: \"8a45f875-20b4-409a-b77a-22305360aa93\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:40:54 crc kubenswrapper[4805]: I1128 15:40:54.962058 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:40:55 crc kubenswrapper[4805]: I1128 15:40:55.384975 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-rqb4d"] Nov 28 15:40:55 crc kubenswrapper[4805]: I1128 15:40:55.524024 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" event={"ID":"8a45f875-20b4-409a-b77a-22305360aa93","Type":"ContainerStarted","Data":"dba480ab6efd2edc709044b8ae36daaf06437513f6bb114c53fae0becf2fdda8"} Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.225142 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lz4"] Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.227641 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.242968 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lz4"] Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.249772 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-catalog-content\") pod \"redhat-marketplace-c2lz4\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.249815 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-utilities\") pod \"redhat-marketplace-c2lz4\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.249878 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq7qf\" (UniqueName: \"kubernetes.io/projected/ad52828c-de09-4ab1-9114-4ee328ae6216-kube-api-access-jq7qf\") pod \"redhat-marketplace-c2lz4\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.293113 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-g55jh"] Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.294201 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.298940 4805 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-klgqz" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.305269 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-g55jh"] Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.351222 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/95ead763-33ad-4a58-905d-7bc10997e6e6-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-g55jh\" (UID: \"95ead763-33ad-4a58-905d-7bc10997e6e6\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.351571 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp9nr\" (UniqueName: \"kubernetes.io/projected/95ead763-33ad-4a58-905d-7bc10997e6e6-kube-api-access-pp9nr\") pod \"cert-manager-cainjector-855d9ccff4-g55jh\" (UID: \"95ead763-33ad-4a58-905d-7bc10997e6e6\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.351776 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-catalog-content\") pod \"redhat-marketplace-c2lz4\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.351906 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-utilities\") pod \"redhat-marketplace-c2lz4\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.352492 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-catalog-content\") pod \"redhat-marketplace-c2lz4\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.352851 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-utilities\") pod \"redhat-marketplace-c2lz4\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.352907 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq7qf\" (UniqueName: \"kubernetes.io/projected/ad52828c-de09-4ab1-9114-4ee328ae6216-kube-api-access-jq7qf\") pod \"redhat-marketplace-c2lz4\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.386378 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq7qf\" (UniqueName: \"kubernetes.io/projected/ad52828c-de09-4ab1-9114-4ee328ae6216-kube-api-access-jq7qf\") pod \"redhat-marketplace-c2lz4\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.454456 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp9nr\" (UniqueName: \"kubernetes.io/projected/95ead763-33ad-4a58-905d-7bc10997e6e6-kube-api-access-pp9nr\") pod \"cert-manager-cainjector-855d9ccff4-g55jh\" (UID: \"95ead763-33ad-4a58-905d-7bc10997e6e6\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.454573 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/95ead763-33ad-4a58-905d-7bc10997e6e6-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-g55jh\" (UID: \"95ead763-33ad-4a58-905d-7bc10997e6e6\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.470264 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/95ead763-33ad-4a58-905d-7bc10997e6e6-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-g55jh\" (UID: \"95ead763-33ad-4a58-905d-7bc10997e6e6\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.472166 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp9nr\" (UniqueName: \"kubernetes.io/projected/95ead763-33ad-4a58-905d-7bc10997e6e6-kube-api-access-pp9nr\") pod \"cert-manager-cainjector-855d9ccff4-g55jh\" (UID: \"95ead763-33ad-4a58-905d-7bc10997e6e6\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.543530 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.610785 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" Nov 28 15:40:56 crc kubenswrapper[4805]: I1128 15:40:56.979022 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lz4"] Nov 28 15:40:57 crc kubenswrapper[4805]: I1128 15:40:57.092863 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-g55jh"] Nov 28 15:40:57 crc kubenswrapper[4805]: W1128 15:40:57.100721 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95ead763_33ad_4a58_905d_7bc10997e6e6.slice/crio-bd78576f7ff19cd32602f01257155cfa6fd05af81d79857a2d0d37014fb29018 WatchSource:0}: Error finding container bd78576f7ff19cd32602f01257155cfa6fd05af81d79857a2d0d37014fb29018: Status 404 returned error can't find the container with id bd78576f7ff19cd32602f01257155cfa6fd05af81d79857a2d0d37014fb29018 Nov 28 15:40:57 crc kubenswrapper[4805]: I1128 15:40:57.537897 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" event={"ID":"95ead763-33ad-4a58-905d-7bc10997e6e6","Type":"ContainerStarted","Data":"bd78576f7ff19cd32602f01257155cfa6fd05af81d79857a2d0d37014fb29018"} Nov 28 15:40:57 crc kubenswrapper[4805]: I1128 15:40:57.540473 4805 generic.go:334] "Generic (PLEG): container finished" podID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerID="6d4637846f613515f1bdc3154e531e6d1e9feb2d1b256fed2b7e769cb66f19fe" exitCode=0 Nov 28 15:40:57 crc kubenswrapper[4805]: I1128 15:40:57.540503 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lz4" event={"ID":"ad52828c-de09-4ab1-9114-4ee328ae6216","Type":"ContainerDied","Data":"6d4637846f613515f1bdc3154e531e6d1e9feb2d1b256fed2b7e769cb66f19fe"} Nov 28 15:40:57 crc kubenswrapper[4805]: I1128 15:40:57.540522 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lz4" event={"ID":"ad52828c-de09-4ab1-9114-4ee328ae6216","Type":"ContainerStarted","Data":"39a2bd07f1296b48dc88d16c8ab6ac9f310e8854c5d52f7486e7ac127d729210"} Nov 28 15:40:58 crc kubenswrapper[4805]: I1128 15:40:58.583255 4805 generic.go:334] "Generic (PLEG): container finished" podID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerID="5d6a3dc8aa6ae4a133c91db010e14aa01b9b80efe2a119f294a5f13eab6e481f" exitCode=0 Nov 28 15:40:58 crc kubenswrapper[4805]: I1128 15:40:58.583587 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lz4" event={"ID":"ad52828c-de09-4ab1-9114-4ee328ae6216","Type":"ContainerDied","Data":"5d6a3dc8aa6ae4a133c91db010e14aa01b9b80efe2a119f294a5f13eab6e481f"} Nov 28 15:41:03 crc kubenswrapper[4805]: I1128 15:41:03.616928 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" event={"ID":"95ead763-33ad-4a58-905d-7bc10997e6e6","Type":"ContainerStarted","Data":"894f67c1b18e4d1aee2b61ee7b5b9c7a5cab0720258825f63e59ef96dd9cf36b"} Nov 28 15:41:03 crc kubenswrapper[4805]: I1128 15:41:03.619204 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" event={"ID":"8a45f875-20b4-409a-b77a-22305360aa93","Type":"ContainerStarted","Data":"d7bb2a624dfc83afedfdc8502d26d18a99cdf7205834d95fc88f67c272b11408"} Nov 28 15:41:03 crc kubenswrapper[4805]: I1128 15:41:03.619575 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:41:03 crc kubenswrapper[4805]: I1128 15:41:03.621876 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lz4" event={"ID":"ad52828c-de09-4ab1-9114-4ee328ae6216","Type":"ContainerStarted","Data":"1b11d0779ac68845ea0122320ffa9d8fb12700f835f3cb41b1fe21d0dc7a0888"} Nov 28 15:41:03 crc kubenswrapper[4805]: I1128 15:41:03.639964 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-g55jh" podStartSLOduration=1.562055846 podStartE2EDuration="7.639942734s" podCreationTimestamp="2025-11-28 15:40:56 +0000 UTC" firstStartedPulling="2025-11-28 15:40:57.102843582 +0000 UTC m=+884.152634893" lastFinishedPulling="2025-11-28 15:41:03.18073046 +0000 UTC m=+890.230521781" observedRunningTime="2025-11-28 15:41:03.633877863 +0000 UTC m=+890.683669164" watchObservedRunningTime="2025-11-28 15:41:03.639942734 +0000 UTC m=+890.689734045" Nov 28 15:41:03 crc kubenswrapper[4805]: I1128 15:41:03.657396 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" podStartSLOduration=1.809645446 podStartE2EDuration="9.657375285s" podCreationTimestamp="2025-11-28 15:40:54 +0000 UTC" firstStartedPulling="2025-11-28 15:40:55.391245704 +0000 UTC m=+882.441037015" lastFinishedPulling="2025-11-28 15:41:03.238975543 +0000 UTC m=+890.288766854" observedRunningTime="2025-11-28 15:41:03.654124279 +0000 UTC m=+890.703915600" watchObservedRunningTime="2025-11-28 15:41:03.657375285 +0000 UTC m=+890.707166596" Nov 28 15:41:03 crc kubenswrapper[4805]: I1128 15:41:03.680406 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c2lz4" podStartSLOduration=2.04239026 podStartE2EDuration="7.680388855s" podCreationTimestamp="2025-11-28 15:40:56 +0000 UTC" firstStartedPulling="2025-11-28 15:40:57.542026946 +0000 UTC m=+884.591818257" lastFinishedPulling="2025-11-28 15:41:03.180025531 +0000 UTC m=+890.229816852" observedRunningTime="2025-11-28 15:41:03.674778226 +0000 UTC m=+890.724569537" watchObservedRunningTime="2025-11-28 15:41:03.680388855 +0000 UTC m=+890.730180166" Nov 28 15:41:06 crc kubenswrapper[4805]: I1128 15:41:06.544046 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:41:06 crc kubenswrapper[4805]: I1128 15:41:06.544668 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:41:06 crc kubenswrapper[4805]: I1128 15:41:06.586137 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:41:09 crc kubenswrapper[4805]: I1128 15:41:09.965012 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-rqb4d" Nov 28 15:41:11 crc kubenswrapper[4805]: I1128 15:41:11.060202 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:41:11 crc kubenswrapper[4805]: I1128 15:41:11.060274 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:41:16 crc kubenswrapper[4805]: I1128 15:41:16.621223 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:41:16 crc kubenswrapper[4805]: I1128 15:41:16.680905 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lz4"] Nov 28 15:41:16 crc kubenswrapper[4805]: I1128 15:41:16.732567 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c2lz4" podUID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerName="registry-server" containerID="cri-o://1b11d0779ac68845ea0122320ffa9d8fb12700f835f3cb41b1fe21d0dc7a0888" gracePeriod=2 Nov 28 15:41:17 crc kubenswrapper[4805]: I1128 15:41:17.739307 4805 generic.go:334] "Generic (PLEG): container finished" podID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerID="1b11d0779ac68845ea0122320ffa9d8fb12700f835f3cb41b1fe21d0dc7a0888" exitCode=0 Nov 28 15:41:17 crc kubenswrapper[4805]: I1128 15:41:17.739375 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lz4" event={"ID":"ad52828c-de09-4ab1-9114-4ee328ae6216","Type":"ContainerDied","Data":"1b11d0779ac68845ea0122320ffa9d8fb12700f835f3cb41b1fe21d0dc7a0888"} Nov 28 15:41:18 crc kubenswrapper[4805]: I1128 15:41:18.827286 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:41:18 crc kubenswrapper[4805]: I1128 15:41:18.944403 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-catalog-content\") pod \"ad52828c-de09-4ab1-9114-4ee328ae6216\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " Nov 28 15:41:18 crc kubenswrapper[4805]: I1128 15:41:18.944449 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-utilities\") pod \"ad52828c-de09-4ab1-9114-4ee328ae6216\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " Nov 28 15:41:18 crc kubenswrapper[4805]: I1128 15:41:18.944480 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq7qf\" (UniqueName: \"kubernetes.io/projected/ad52828c-de09-4ab1-9114-4ee328ae6216-kube-api-access-jq7qf\") pod \"ad52828c-de09-4ab1-9114-4ee328ae6216\" (UID: \"ad52828c-de09-4ab1-9114-4ee328ae6216\") " Nov 28 15:41:18 crc kubenswrapper[4805]: I1128 15:41:18.945841 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-utilities" (OuterVolumeSpecName: "utilities") pod "ad52828c-de09-4ab1-9114-4ee328ae6216" (UID: "ad52828c-de09-4ab1-9114-4ee328ae6216"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:41:18 crc kubenswrapper[4805]: I1128 15:41:18.953599 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad52828c-de09-4ab1-9114-4ee328ae6216-kube-api-access-jq7qf" (OuterVolumeSpecName: "kube-api-access-jq7qf") pod "ad52828c-de09-4ab1-9114-4ee328ae6216" (UID: "ad52828c-de09-4ab1-9114-4ee328ae6216"). InnerVolumeSpecName "kube-api-access-jq7qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:41:18 crc kubenswrapper[4805]: I1128 15:41:18.972048 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad52828c-de09-4ab1-9114-4ee328ae6216" (UID: "ad52828c-de09-4ab1-9114-4ee328ae6216"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.047169 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.047206 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq7qf\" (UniqueName: \"kubernetes.io/projected/ad52828c-de09-4ab1-9114-4ee328ae6216-kube-api-access-jq7qf\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.047215 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad52828c-de09-4ab1-9114-4ee328ae6216-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.754224 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lz4" event={"ID":"ad52828c-de09-4ab1-9114-4ee328ae6216","Type":"ContainerDied","Data":"39a2bd07f1296b48dc88d16c8ab6ac9f310e8854c5d52f7486e7ac127d729210"} Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.754293 4805 scope.go:117] "RemoveContainer" containerID="1b11d0779ac68845ea0122320ffa9d8fb12700f835f3cb41b1fe21d0dc7a0888" Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.754303 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lz4" Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.774542 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lz4"] Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.779747 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lz4"] Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.780724 4805 scope.go:117] "RemoveContainer" containerID="5d6a3dc8aa6ae4a133c91db010e14aa01b9b80efe2a119f294a5f13eab6e481f" Nov 28 15:41:19 crc kubenswrapper[4805]: I1128 15:41:19.801139 4805 scope.go:117] "RemoveContainer" containerID="6d4637846f613515f1bdc3154e531e6d1e9feb2d1b256fed2b7e769cb66f19fe" Nov 28 15:41:21 crc kubenswrapper[4805]: I1128 15:41:21.217972 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad52828c-de09-4ab1-9114-4ee328ae6216" path="/var/lib/kubelet/pods/ad52828c-de09-4ab1-9114-4ee328ae6216/volumes" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.612773 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-65t7k"] Nov 28 15:41:22 crc kubenswrapper[4805]: E1128 15:41:22.613436 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerName="extract-utilities" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.613455 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerName="extract-utilities" Nov 28 15:41:22 crc kubenswrapper[4805]: E1128 15:41:22.613479 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerName="extract-content" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.613489 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerName="extract-content" Nov 28 15:41:22 crc kubenswrapper[4805]: E1128 15:41:22.613540 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerName="registry-server" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.613553 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerName="registry-server" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.613723 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad52828c-de09-4ab1-9114-4ee328ae6216" containerName="registry-server" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.614303 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-65t7k" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.616440 4805 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-xp8x8" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.625088 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-65t7k"] Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.794483 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1228d1c5-60b6-4b46-8f2f-f163ff0dea96-bound-sa-token\") pod \"cert-manager-86cb77c54b-65t7k\" (UID: \"1228d1c5-60b6-4b46-8f2f-f163ff0dea96\") " pod="cert-manager/cert-manager-86cb77c54b-65t7k" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.794794 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74vdn\" (UniqueName: \"kubernetes.io/projected/1228d1c5-60b6-4b46-8f2f-f163ff0dea96-kube-api-access-74vdn\") pod \"cert-manager-86cb77c54b-65t7k\" (UID: \"1228d1c5-60b6-4b46-8f2f-f163ff0dea96\") " pod="cert-manager/cert-manager-86cb77c54b-65t7k" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.895756 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74vdn\" (UniqueName: \"kubernetes.io/projected/1228d1c5-60b6-4b46-8f2f-f163ff0dea96-kube-api-access-74vdn\") pod \"cert-manager-86cb77c54b-65t7k\" (UID: \"1228d1c5-60b6-4b46-8f2f-f163ff0dea96\") " pod="cert-manager/cert-manager-86cb77c54b-65t7k" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.895837 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1228d1c5-60b6-4b46-8f2f-f163ff0dea96-bound-sa-token\") pod \"cert-manager-86cb77c54b-65t7k\" (UID: \"1228d1c5-60b6-4b46-8f2f-f163ff0dea96\") " pod="cert-manager/cert-manager-86cb77c54b-65t7k" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.929293 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1228d1c5-60b6-4b46-8f2f-f163ff0dea96-bound-sa-token\") pod \"cert-manager-86cb77c54b-65t7k\" (UID: \"1228d1c5-60b6-4b46-8f2f-f163ff0dea96\") " pod="cert-manager/cert-manager-86cb77c54b-65t7k" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.930303 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74vdn\" (UniqueName: \"kubernetes.io/projected/1228d1c5-60b6-4b46-8f2f-f163ff0dea96-kube-api-access-74vdn\") pod \"cert-manager-86cb77c54b-65t7k\" (UID: \"1228d1c5-60b6-4b46-8f2f-f163ff0dea96\") " pod="cert-manager/cert-manager-86cb77c54b-65t7k" Nov 28 15:41:22 crc kubenswrapper[4805]: I1128 15:41:22.936598 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-65t7k" Nov 28 15:41:23 crc kubenswrapper[4805]: I1128 15:41:23.382996 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-65t7k"] Nov 28 15:41:23 crc kubenswrapper[4805]: W1128 15:41:23.387470 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1228d1c5_60b6_4b46_8f2f_f163ff0dea96.slice/crio-471957bdbc7ec74344189175632eded4cd5d10d9f7f4413fe23a25e287bc6934 WatchSource:0}: Error finding container 471957bdbc7ec74344189175632eded4cd5d10d9f7f4413fe23a25e287bc6934: Status 404 returned error can't find the container with id 471957bdbc7ec74344189175632eded4cd5d10d9f7f4413fe23a25e287bc6934 Nov 28 15:41:23 crc kubenswrapper[4805]: I1128 15:41:23.779614 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-65t7k" event={"ID":"1228d1c5-60b6-4b46-8f2f-f163ff0dea96","Type":"ContainerStarted","Data":"471957bdbc7ec74344189175632eded4cd5d10d9f7f4413fe23a25e287bc6934"} Nov 28 15:41:25 crc kubenswrapper[4805]: I1128 15:41:25.792089 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-65t7k" event={"ID":"1228d1c5-60b6-4b46-8f2f-f163ff0dea96","Type":"ContainerStarted","Data":"193ec5eac8d84f5528b2ec2467344af77f04d78c6cc1406e76ece9b5f30c3720"} Nov 28 15:41:25 crc kubenswrapper[4805]: I1128 15:41:25.814164 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-65t7k" podStartSLOduration=3.8141458630000002 podStartE2EDuration="3.814145863s" podCreationTimestamp="2025-11-28 15:41:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:41:25.809428449 +0000 UTC m=+912.859219760" watchObservedRunningTime="2025-11-28 15:41:25.814145863 +0000 UTC m=+912.863937174" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.490937 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gfwh4"] Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.493156 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.499861 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gfwh4"] Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.530237 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctrh4\" (UniqueName: \"kubernetes.io/projected/c71f61c2-01ca-424b-9298-a37c38c0e687-kube-api-access-ctrh4\") pod \"community-operators-gfwh4\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.530338 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-utilities\") pod \"community-operators-gfwh4\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.530390 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-catalog-content\") pod \"community-operators-gfwh4\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.631197 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-utilities\") pod \"community-operators-gfwh4\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.631264 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-catalog-content\") pod \"community-operators-gfwh4\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.631309 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctrh4\" (UniqueName: \"kubernetes.io/projected/c71f61c2-01ca-424b-9298-a37c38c0e687-kube-api-access-ctrh4\") pod \"community-operators-gfwh4\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.631777 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-utilities\") pod \"community-operators-gfwh4\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.631797 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-catalog-content\") pod \"community-operators-gfwh4\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.663106 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctrh4\" (UniqueName: \"kubernetes.io/projected/c71f61c2-01ca-424b-9298-a37c38c0e687-kube-api-access-ctrh4\") pod \"community-operators-gfwh4\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:33 crc kubenswrapper[4805]: I1128 15:41:33.823273 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:34 crc kubenswrapper[4805]: I1128 15:41:34.356592 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gfwh4"] Nov 28 15:41:34 crc kubenswrapper[4805]: I1128 15:41:34.876291 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfwh4" event={"ID":"c71f61c2-01ca-424b-9298-a37c38c0e687","Type":"ContainerStarted","Data":"dd605aca42733c4929303253da8b179e2a9b128d551f7cc526df083a99a4cd4d"} Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.088570 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-lz8tb"] Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.089711 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lz8tb" Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.091687 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-wbnwr" Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.091839 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.091885 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.099176 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lz8tb"] Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.193236 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtzdt\" (UniqueName: \"kubernetes.io/projected/e163631e-4442-4a96-a1e2-7b0ea21fd517-kube-api-access-jtzdt\") pod \"openstack-operator-index-lz8tb\" (UID: \"e163631e-4442-4a96-a1e2-7b0ea21fd517\") " pod="openstack-operators/openstack-operator-index-lz8tb" Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.294751 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtzdt\" (UniqueName: \"kubernetes.io/projected/e163631e-4442-4a96-a1e2-7b0ea21fd517-kube-api-access-jtzdt\") pod \"openstack-operator-index-lz8tb\" (UID: \"e163631e-4442-4a96-a1e2-7b0ea21fd517\") " pod="openstack-operators/openstack-operator-index-lz8tb" Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.317323 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtzdt\" (UniqueName: \"kubernetes.io/projected/e163631e-4442-4a96-a1e2-7b0ea21fd517-kube-api-access-jtzdt\") pod \"openstack-operator-index-lz8tb\" (UID: \"e163631e-4442-4a96-a1e2-7b0ea21fd517\") " pod="openstack-operators/openstack-operator-index-lz8tb" Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.405023 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lz8tb" Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.694338 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lz8tb"] Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.901925 4805 generic.go:334] "Generic (PLEG): container finished" podID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerID="32d027b8e872cc0576157c8ccf4cffa11b880c381b723c8b0c15b9505550afc4" exitCode=0 Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.901994 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfwh4" event={"ID":"c71f61c2-01ca-424b-9298-a37c38c0e687","Type":"ContainerDied","Data":"32d027b8e872cc0576157c8ccf4cffa11b880c381b723c8b0c15b9505550afc4"} Nov 28 15:41:37 crc kubenswrapper[4805]: I1128 15:41:37.903297 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lz8tb" event={"ID":"e163631e-4442-4a96-a1e2-7b0ea21fd517","Type":"ContainerStarted","Data":"419c029a708b928ae7069b925239fb143429c89301eca497c17ca07303ac4ed4"} Nov 28 15:41:38 crc kubenswrapper[4805]: I1128 15:41:38.909952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lz8tb" event={"ID":"e163631e-4442-4a96-a1e2-7b0ea21fd517","Type":"ContainerStarted","Data":"e38cbe900c203e3e171e9605ad38d07c4bd2e086b1e8666a7afbea263b693e5c"} Nov 28 15:41:38 crc kubenswrapper[4805]: I1128 15:41:38.911936 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfwh4" event={"ID":"c71f61c2-01ca-424b-9298-a37c38c0e687","Type":"ContainerStarted","Data":"5dcfa79e3056bcead728c317fef53686207b765501a5fb7e0d4f25d0b8edd449"} Nov 28 15:41:38 crc kubenswrapper[4805]: I1128 15:41:38.937506 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-lz8tb" podStartSLOduration=1.163301078 podStartE2EDuration="1.937489576s" podCreationTimestamp="2025-11-28 15:41:37 +0000 UTC" firstStartedPulling="2025-11-28 15:41:37.705902423 +0000 UTC m=+924.755693744" lastFinishedPulling="2025-11-28 15:41:38.480090941 +0000 UTC m=+925.529882242" observedRunningTime="2025-11-28 15:41:38.936297635 +0000 UTC m=+925.986088956" watchObservedRunningTime="2025-11-28 15:41:38.937489576 +0000 UTC m=+925.987280887" Nov 28 15:41:39 crc kubenswrapper[4805]: I1128 15:41:39.923154 4805 generic.go:334] "Generic (PLEG): container finished" podID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerID="5dcfa79e3056bcead728c317fef53686207b765501a5fb7e0d4f25d0b8edd449" exitCode=0 Nov 28 15:41:39 crc kubenswrapper[4805]: I1128 15:41:39.923232 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfwh4" event={"ID":"c71f61c2-01ca-424b-9298-a37c38c0e687","Type":"ContainerDied","Data":"5dcfa79e3056bcead728c317fef53686207b765501a5fb7e0d4f25d0b8edd449"} Nov 28 15:41:40 crc kubenswrapper[4805]: I1128 15:41:40.938077 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfwh4" event={"ID":"c71f61c2-01ca-424b-9298-a37c38c0e687","Type":"ContainerStarted","Data":"aa23e5b6e1d0bff09a5fab07983e8cdff900b8efff7108a269eb45026996547c"} Nov 28 15:41:40 crc kubenswrapper[4805]: I1128 15:41:40.961267 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gfwh4" podStartSLOduration=5.215933863 podStartE2EDuration="7.961244434s" podCreationTimestamp="2025-11-28 15:41:33 +0000 UTC" firstStartedPulling="2025-11-28 15:41:37.905315756 +0000 UTC m=+924.955107107" lastFinishedPulling="2025-11-28 15:41:40.650626367 +0000 UTC m=+927.700417678" observedRunningTime="2025-11-28 15:41:40.955714477 +0000 UTC m=+928.005505788" watchObservedRunningTime="2025-11-28 15:41:40.961244434 +0000 UTC m=+928.011035765" Nov 28 15:41:41 crc kubenswrapper[4805]: I1128 15:41:41.060340 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:41:41 crc kubenswrapper[4805]: I1128 15:41:41.060409 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.070699 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7mkbg"] Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.072340 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.094694 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mkbg"] Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.161337 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxm82\" (UniqueName: \"kubernetes.io/projected/2bef090e-f1af-4926-b364-d8414d8835a4-kube-api-access-jxm82\") pod \"certified-operators-7mkbg\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.161409 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-utilities\") pod \"certified-operators-7mkbg\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.161430 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-catalog-content\") pod \"certified-operators-7mkbg\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.262817 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxm82\" (UniqueName: \"kubernetes.io/projected/2bef090e-f1af-4926-b364-d8414d8835a4-kube-api-access-jxm82\") pod \"certified-operators-7mkbg\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.262895 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-utilities\") pod \"certified-operators-7mkbg\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.262944 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-catalog-content\") pod \"certified-operators-7mkbg\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.264491 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-utilities\") pod \"certified-operators-7mkbg\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.264820 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-catalog-content\") pod \"certified-operators-7mkbg\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.284747 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxm82\" (UniqueName: \"kubernetes.io/projected/2bef090e-f1af-4926-b364-d8414d8835a4-kube-api-access-jxm82\") pod \"certified-operators-7mkbg\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.391636 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.831056 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mkbg"] Nov 28 15:41:42 crc kubenswrapper[4805]: I1128 15:41:42.949870 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkbg" event={"ID":"2bef090e-f1af-4926-b364-d8414d8835a4","Type":"ContainerStarted","Data":"dca2281cd1b2d69e27296b2161c69ae062b8e34b112f0ba90637e54c318d6c0a"} Nov 28 15:41:43 crc kubenswrapper[4805]: I1128 15:41:43.823843 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:43 crc kubenswrapper[4805]: I1128 15:41:43.824101 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:43 crc kubenswrapper[4805]: I1128 15:41:43.877098 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:43 crc kubenswrapper[4805]: I1128 15:41:43.961870 4805 generic.go:334] "Generic (PLEG): container finished" podID="2bef090e-f1af-4926-b364-d8414d8835a4" containerID="b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21" exitCode=0 Nov 28 15:41:43 crc kubenswrapper[4805]: I1128 15:41:43.961925 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkbg" event={"ID":"2bef090e-f1af-4926-b364-d8414d8835a4","Type":"ContainerDied","Data":"b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21"} Nov 28 15:41:45 crc kubenswrapper[4805]: I1128 15:41:45.976025 4805 generic.go:334] "Generic (PLEG): container finished" podID="2bef090e-f1af-4926-b364-d8414d8835a4" containerID="4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2" exitCode=0 Nov 28 15:41:45 crc kubenswrapper[4805]: I1128 15:41:45.976080 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkbg" event={"ID":"2bef090e-f1af-4926-b364-d8414d8835a4","Type":"ContainerDied","Data":"4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2"} Nov 28 15:41:46 crc kubenswrapper[4805]: I1128 15:41:46.984064 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkbg" event={"ID":"2bef090e-f1af-4926-b364-d8414d8835a4","Type":"ContainerStarted","Data":"c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b"} Nov 28 15:41:47 crc kubenswrapper[4805]: I1128 15:41:47.011114 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7mkbg" podStartSLOduration=2.433461299 podStartE2EDuration="5.011091867s" podCreationTimestamp="2025-11-28 15:41:42 +0000 UTC" firstStartedPulling="2025-11-28 15:41:43.963685895 +0000 UTC m=+931.013477226" lastFinishedPulling="2025-11-28 15:41:46.541316483 +0000 UTC m=+933.591107794" observedRunningTime="2025-11-28 15:41:47.007508503 +0000 UTC m=+934.057299834" watchObservedRunningTime="2025-11-28 15:41:47.011091867 +0000 UTC m=+934.060883188" Nov 28 15:41:47 crc kubenswrapper[4805]: I1128 15:41:47.405737 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-lz8tb" Nov 28 15:41:47 crc kubenswrapper[4805]: I1128 15:41:47.405816 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-lz8tb" Nov 28 15:41:47 crc kubenswrapper[4805]: I1128 15:41:47.462706 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-lz8tb" Nov 28 15:41:48 crc kubenswrapper[4805]: I1128 15:41:48.041221 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-lz8tb" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.511076 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd"] Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.512334 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.514959 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-djwmt" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.523394 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd"] Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.574147 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-bundle\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.574190 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpznh\" (UniqueName: \"kubernetes.io/projected/7e48a537-6a16-4037-96aa-242276b036a8-kube-api-access-bpznh\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.574247 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-util\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.675524 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-bundle\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.675563 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpznh\" (UniqueName: \"kubernetes.io/projected/7e48a537-6a16-4037-96aa-242276b036a8-kube-api-access-bpznh\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.675622 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-util\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.676034 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-util\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.676422 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-bundle\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.693225 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpznh\" (UniqueName: \"kubernetes.io/projected/7e48a537-6a16-4037-96aa-242276b036a8-kube-api-access-bpznh\") pod \"5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:49 crc kubenswrapper[4805]: I1128 15:41:49.825530 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:50 crc kubenswrapper[4805]: I1128 15:41:50.262459 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd"] Nov 28 15:41:50 crc kubenswrapper[4805]: W1128 15:41:50.283335 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e48a537_6a16_4037_96aa_242276b036a8.slice/crio-787c8096e114a7d34f97e198ec6bbfa6f3cf1e547e942226e1d8a8aa7a1be039 WatchSource:0}: Error finding container 787c8096e114a7d34f97e198ec6bbfa6f3cf1e547e942226e1d8a8aa7a1be039: Status 404 returned error can't find the container with id 787c8096e114a7d34f97e198ec6bbfa6f3cf1e547e942226e1d8a8aa7a1be039 Nov 28 15:41:51 crc kubenswrapper[4805]: I1128 15:41:51.029898 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" event={"ID":"7e48a537-6a16-4037-96aa-242276b036a8","Type":"ContainerStarted","Data":"6d8957a04cdc270734d2078f795c2d4d7b4601a1a2a3cb0fa01028dae8188ad1"} Nov 28 15:41:51 crc kubenswrapper[4805]: I1128 15:41:51.030183 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" event={"ID":"7e48a537-6a16-4037-96aa-242276b036a8","Type":"ContainerStarted","Data":"787c8096e114a7d34f97e198ec6bbfa6f3cf1e547e942226e1d8a8aa7a1be039"} Nov 28 15:41:52 crc kubenswrapper[4805]: I1128 15:41:52.037575 4805 generic.go:334] "Generic (PLEG): container finished" podID="7e48a537-6a16-4037-96aa-242276b036a8" containerID="6d8957a04cdc270734d2078f795c2d4d7b4601a1a2a3cb0fa01028dae8188ad1" exitCode=0 Nov 28 15:41:52 crc kubenswrapper[4805]: I1128 15:41:52.037637 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" event={"ID":"7e48a537-6a16-4037-96aa-242276b036a8","Type":"ContainerDied","Data":"6d8957a04cdc270734d2078f795c2d4d7b4601a1a2a3cb0fa01028dae8188ad1"} Nov 28 15:41:52 crc kubenswrapper[4805]: I1128 15:41:52.392532 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:52 crc kubenswrapper[4805]: I1128 15:41:52.392802 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:52 crc kubenswrapper[4805]: I1128 15:41:52.441132 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:53 crc kubenswrapper[4805]: I1128 15:41:53.046794 4805 generic.go:334] "Generic (PLEG): container finished" podID="7e48a537-6a16-4037-96aa-242276b036a8" containerID="11555ee7bbfe6e9309a76a48c3fb4ada7dd4ab0b6bf280e522c3ec9f3f334f52" exitCode=0 Nov 28 15:41:53 crc kubenswrapper[4805]: I1128 15:41:53.046892 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" event={"ID":"7e48a537-6a16-4037-96aa-242276b036a8","Type":"ContainerDied","Data":"11555ee7bbfe6e9309a76a48c3fb4ada7dd4ab0b6bf280e522c3ec9f3f334f52"} Nov 28 15:41:53 crc kubenswrapper[4805]: I1128 15:41:53.104316 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:53 crc kubenswrapper[4805]: I1128 15:41:53.871014 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mkbg"] Nov 28 15:41:53 crc kubenswrapper[4805]: I1128 15:41:53.875563 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:54 crc kubenswrapper[4805]: I1128 15:41:54.057177 4805 generic.go:334] "Generic (PLEG): container finished" podID="7e48a537-6a16-4037-96aa-242276b036a8" containerID="33d625abd2bf438d86795c1342346d5a404896516624666f4bcbad8a430dc110" exitCode=0 Nov 28 15:41:54 crc kubenswrapper[4805]: I1128 15:41:54.057300 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" event={"ID":"7e48a537-6a16-4037-96aa-242276b036a8","Type":"ContainerDied","Data":"33d625abd2bf438d86795c1342346d5a404896516624666f4bcbad8a430dc110"} Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.064385 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7mkbg" podUID="2bef090e-f1af-4926-b364-d8414d8835a4" containerName="registry-server" containerID="cri-o://c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b" gracePeriod=2 Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.471525 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.475517 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:55 crc kubenswrapper[4805]: E1128 15:41:55.476763 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bef090e_f1af_4926_b364_d8414d8835a4.slice/crio-conmon-c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b.scope\": RecentStats: unable to find data in memory cache]" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.600872 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-catalog-content\") pod \"2bef090e-f1af-4926-b364-d8414d8835a4\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.600923 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-bundle\") pod \"7e48a537-6a16-4037-96aa-242276b036a8\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.600944 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpznh\" (UniqueName: \"kubernetes.io/projected/7e48a537-6a16-4037-96aa-242276b036a8-kube-api-access-bpznh\") pod \"7e48a537-6a16-4037-96aa-242276b036a8\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.600974 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxm82\" (UniqueName: \"kubernetes.io/projected/2bef090e-f1af-4926-b364-d8414d8835a4-kube-api-access-jxm82\") pod \"2bef090e-f1af-4926-b364-d8414d8835a4\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.601034 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-utilities\") pod \"2bef090e-f1af-4926-b364-d8414d8835a4\" (UID: \"2bef090e-f1af-4926-b364-d8414d8835a4\") " Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.601086 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-util\") pod \"7e48a537-6a16-4037-96aa-242276b036a8\" (UID: \"7e48a537-6a16-4037-96aa-242276b036a8\") " Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.601886 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-bundle" (OuterVolumeSpecName: "bundle") pod "7e48a537-6a16-4037-96aa-242276b036a8" (UID: "7e48a537-6a16-4037-96aa-242276b036a8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.602526 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-utilities" (OuterVolumeSpecName: "utilities") pod "2bef090e-f1af-4926-b364-d8414d8835a4" (UID: "2bef090e-f1af-4926-b364-d8414d8835a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.607538 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bef090e-f1af-4926-b364-d8414d8835a4-kube-api-access-jxm82" (OuterVolumeSpecName: "kube-api-access-jxm82") pod "2bef090e-f1af-4926-b364-d8414d8835a4" (UID: "2bef090e-f1af-4926-b364-d8414d8835a4"). InnerVolumeSpecName "kube-api-access-jxm82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.611023 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e48a537-6a16-4037-96aa-242276b036a8-kube-api-access-bpznh" (OuterVolumeSpecName: "kube-api-access-bpznh") pod "7e48a537-6a16-4037-96aa-242276b036a8" (UID: "7e48a537-6a16-4037-96aa-242276b036a8"). InnerVolumeSpecName "kube-api-access-bpznh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.615062 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-util" (OuterVolumeSpecName: "util") pod "7e48a537-6a16-4037-96aa-242276b036a8" (UID: "7e48a537-6a16-4037-96aa-242276b036a8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.652631 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bef090e-f1af-4926-b364-d8414d8835a4" (UID: "2bef090e-f1af-4926-b364-d8414d8835a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.702382 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxm82\" (UniqueName: \"kubernetes.io/projected/2bef090e-f1af-4926-b364-d8414d8835a4-kube-api-access-jxm82\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.702663 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.702742 4805 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-util\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.702812 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bef090e-f1af-4926-b364-d8414d8835a4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.702898 4805 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e48a537-6a16-4037-96aa-242276b036a8-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:55 crc kubenswrapper[4805]: I1128 15:41:55.703006 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpznh\" (UniqueName: \"kubernetes.io/projected/7e48a537-6a16-4037-96aa-242276b036a8-kube-api-access-bpznh\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.078018 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" event={"ID":"7e48a537-6a16-4037-96aa-242276b036a8","Type":"ContainerDied","Data":"787c8096e114a7d34f97e198ec6bbfa6f3cf1e547e942226e1d8a8aa7a1be039"} Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.078083 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="787c8096e114a7d34f97e198ec6bbfa6f3cf1e547e942226e1d8a8aa7a1be039" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.078198 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.083342 4805 generic.go:334] "Generic (PLEG): container finished" podID="2bef090e-f1af-4926-b364-d8414d8835a4" containerID="c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b" exitCode=0 Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.083419 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkbg" event={"ID":"2bef090e-f1af-4926-b364-d8414d8835a4","Type":"ContainerDied","Data":"c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b"} Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.083444 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mkbg" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.083458 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkbg" event={"ID":"2bef090e-f1af-4926-b364-d8414d8835a4","Type":"ContainerDied","Data":"dca2281cd1b2d69e27296b2161c69ae062b8e34b112f0ba90637e54c318d6c0a"} Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.083489 4805 scope.go:117] "RemoveContainer" containerID="c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.111000 4805 scope.go:117] "RemoveContainer" containerID="4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.139931 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mkbg"] Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.143891 4805 scope.go:117] "RemoveContainer" containerID="b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.148376 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7mkbg"] Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.165634 4805 scope.go:117] "RemoveContainer" containerID="c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b" Nov 28 15:41:56 crc kubenswrapper[4805]: E1128 15:41:56.166466 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b\": container with ID starting with c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b not found: ID does not exist" containerID="c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.166507 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b"} err="failed to get container status \"c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b\": rpc error: code = NotFound desc = could not find container \"c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b\": container with ID starting with c5fadb5b447710aca2fa5e1977f58523add2af168be956d1aceb4b003ec8b84b not found: ID does not exist" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.166541 4805 scope.go:117] "RemoveContainer" containerID="4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2" Nov 28 15:41:56 crc kubenswrapper[4805]: E1128 15:41:56.167009 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2\": container with ID starting with 4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2 not found: ID does not exist" containerID="4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.167052 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2"} err="failed to get container status \"4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2\": rpc error: code = NotFound desc = could not find container \"4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2\": container with ID starting with 4e5e231d96e8bbe9af19ce9c7de3a300a33b05677347a1c529a44e80da9199e2 not found: ID does not exist" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.167069 4805 scope.go:117] "RemoveContainer" containerID="b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21" Nov 28 15:41:56 crc kubenswrapper[4805]: E1128 15:41:56.167494 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21\": container with ID starting with b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21 not found: ID does not exist" containerID="b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21" Nov 28 15:41:56 crc kubenswrapper[4805]: I1128 15:41:56.167526 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21"} err="failed to get container status \"b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21\": rpc error: code = NotFound desc = could not find container \"b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21\": container with ID starting with b5630bcaa1349275d4633f0c709533a1f01ccc775a4430e8eca2279f27185f21 not found: ID does not exist" Nov 28 15:41:57 crc kubenswrapper[4805]: I1128 15:41:57.219281 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bef090e-f1af-4926-b364-d8414d8835a4" path="/var/lib/kubelet/pods/2bef090e-f1af-4926-b364-d8414d8835a4/volumes" Nov 28 15:41:57 crc kubenswrapper[4805]: I1128 15:41:57.670553 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gfwh4"] Nov 28 15:41:57 crc kubenswrapper[4805]: I1128 15:41:57.670983 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gfwh4" podUID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerName="registry-server" containerID="cri-o://aa23e5b6e1d0bff09a5fab07983e8cdff900b8efff7108a269eb45026996547c" gracePeriod=2 Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.104869 4805 generic.go:334] "Generic (PLEG): container finished" podID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerID="aa23e5b6e1d0bff09a5fab07983e8cdff900b8efff7108a269eb45026996547c" exitCode=0 Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.104947 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfwh4" event={"ID":"c71f61c2-01ca-424b-9298-a37c38c0e687","Type":"ContainerDied","Data":"aa23e5b6e1d0bff09a5fab07983e8cdff900b8efff7108a269eb45026996547c"} Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.105296 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfwh4" event={"ID":"c71f61c2-01ca-424b-9298-a37c38c0e687","Type":"ContainerDied","Data":"dd605aca42733c4929303253da8b179e2a9b128d551f7cc526df083a99a4cd4d"} Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.105319 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd605aca42733c4929303253da8b179e2a9b128d551f7cc526df083a99a4cd4d" Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.113223 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.242525 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-utilities\") pod \"c71f61c2-01ca-424b-9298-a37c38c0e687\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.242597 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctrh4\" (UniqueName: \"kubernetes.io/projected/c71f61c2-01ca-424b-9298-a37c38c0e687-kube-api-access-ctrh4\") pod \"c71f61c2-01ca-424b-9298-a37c38c0e687\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.242670 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-catalog-content\") pod \"c71f61c2-01ca-424b-9298-a37c38c0e687\" (UID: \"c71f61c2-01ca-424b-9298-a37c38c0e687\") " Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.252339 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-utilities" (OuterVolumeSpecName: "utilities") pod "c71f61c2-01ca-424b-9298-a37c38c0e687" (UID: "c71f61c2-01ca-424b-9298-a37c38c0e687"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.266610 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c71f61c2-01ca-424b-9298-a37c38c0e687-kube-api-access-ctrh4" (OuterVolumeSpecName: "kube-api-access-ctrh4") pod "c71f61c2-01ca-424b-9298-a37c38c0e687" (UID: "c71f61c2-01ca-424b-9298-a37c38c0e687"). InnerVolumeSpecName "kube-api-access-ctrh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.299789 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c71f61c2-01ca-424b-9298-a37c38c0e687" (UID: "c71f61c2-01ca-424b-9298-a37c38c0e687"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.344644 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.344677 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f61c2-01ca-424b-9298-a37c38c0e687-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:58 crc kubenswrapper[4805]: I1128 15:41:58.344686 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctrh4\" (UniqueName: \"kubernetes.io/projected/c71f61c2-01ca-424b-9298-a37c38c0e687-kube-api-access-ctrh4\") on node \"crc\" DevicePath \"\"" Nov 28 15:41:59 crc kubenswrapper[4805]: I1128 15:41:59.112471 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gfwh4" Nov 28 15:41:59 crc kubenswrapper[4805]: I1128 15:41:59.156576 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gfwh4"] Nov 28 15:41:59 crc kubenswrapper[4805]: I1128 15:41:59.164186 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gfwh4"] Nov 28 15:41:59 crc kubenswrapper[4805]: I1128 15:41:59.221171 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c71f61c2-01ca-424b-9298-a37c38c0e687" path="/var/lib/kubelet/pods/c71f61c2-01ca-424b-9298-a37c38c0e687/volumes" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.053748 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm"] Nov 28 15:42:01 crc kubenswrapper[4805]: E1128 15:42:01.054470 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerName="extract-utilities" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054492 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerName="extract-utilities" Nov 28 15:42:01 crc kubenswrapper[4805]: E1128 15:42:01.054514 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e48a537-6a16-4037-96aa-242276b036a8" containerName="pull" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054525 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e48a537-6a16-4037-96aa-242276b036a8" containerName="pull" Nov 28 15:42:01 crc kubenswrapper[4805]: E1128 15:42:01.054560 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e48a537-6a16-4037-96aa-242276b036a8" containerName="util" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054571 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e48a537-6a16-4037-96aa-242276b036a8" containerName="util" Nov 28 15:42:01 crc kubenswrapper[4805]: E1128 15:42:01.054585 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bef090e-f1af-4926-b364-d8414d8835a4" containerName="extract-utilities" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054599 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bef090e-f1af-4926-b364-d8414d8835a4" containerName="extract-utilities" Nov 28 15:42:01 crc kubenswrapper[4805]: E1128 15:42:01.054622 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e48a537-6a16-4037-96aa-242276b036a8" containerName="extract" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054633 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e48a537-6a16-4037-96aa-242276b036a8" containerName="extract" Nov 28 15:42:01 crc kubenswrapper[4805]: E1128 15:42:01.054654 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerName="registry-server" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054665 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerName="registry-server" Nov 28 15:42:01 crc kubenswrapper[4805]: E1128 15:42:01.054678 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bef090e-f1af-4926-b364-d8414d8835a4" containerName="extract-content" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054687 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bef090e-f1af-4926-b364-d8414d8835a4" containerName="extract-content" Nov 28 15:42:01 crc kubenswrapper[4805]: E1128 15:42:01.054705 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerName="extract-content" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054716 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerName="extract-content" Nov 28 15:42:01 crc kubenswrapper[4805]: E1128 15:42:01.054736 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bef090e-f1af-4926-b364-d8414d8835a4" containerName="registry-server" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054747 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bef090e-f1af-4926-b364-d8414d8835a4" containerName="registry-server" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054948 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c71f61c2-01ca-424b-9298-a37c38c0e687" containerName="registry-server" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.054977 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e48a537-6a16-4037-96aa-242276b036a8" containerName="extract" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.055008 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bef090e-f1af-4926-b364-d8414d8835a4" containerName="registry-server" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.055685 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.059111 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-rvgcs" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.124192 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm"] Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.183546 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq6cx\" (UniqueName: \"kubernetes.io/projected/af063d23-6938-4b22-b491-60e327e03f2b-kube-api-access-sq6cx\") pod \"openstack-operator-controller-operator-848c47db98-7l2qm\" (UID: \"af063d23-6938-4b22-b491-60e327e03f2b\") " pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.284689 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq6cx\" (UniqueName: \"kubernetes.io/projected/af063d23-6938-4b22-b491-60e327e03f2b-kube-api-access-sq6cx\") pod \"openstack-operator-controller-operator-848c47db98-7l2qm\" (UID: \"af063d23-6938-4b22-b491-60e327e03f2b\") " pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.301744 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq6cx\" (UniqueName: \"kubernetes.io/projected/af063d23-6938-4b22-b491-60e327e03f2b-kube-api-access-sq6cx\") pod \"openstack-operator-controller-operator-848c47db98-7l2qm\" (UID: \"af063d23-6938-4b22-b491-60e327e03f2b\") " pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.371108 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" Nov 28 15:42:01 crc kubenswrapper[4805]: I1128 15:42:01.611387 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm"] Nov 28 15:42:02 crc kubenswrapper[4805]: I1128 15:42:02.145703 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" event={"ID":"af063d23-6938-4b22-b491-60e327e03f2b","Type":"ContainerStarted","Data":"15bd60d21daac9550fb87e61911d76f4138360960a8997eb13f790c22666ba06"} Nov 28 15:42:07 crc kubenswrapper[4805]: I1128 15:42:07.186248 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" event={"ID":"af063d23-6938-4b22-b491-60e327e03f2b","Type":"ContainerStarted","Data":"7eecce4df347363d06c965ca7b122372e7bcef395937ebaadcc3c75da5241f7c"} Nov 28 15:42:07 crc kubenswrapper[4805]: I1128 15:42:07.186884 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" Nov 28 15:42:11 crc kubenswrapper[4805]: I1128 15:42:11.060427 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:42:11 crc kubenswrapper[4805]: I1128 15:42:11.060822 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:42:11 crc kubenswrapper[4805]: I1128 15:42:11.060884 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:42:11 crc kubenswrapper[4805]: I1128 15:42:11.061749 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9401a6879e1fcc43d0b27657309da77bf635d82861975cc53606a87548ec258e"} pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:42:11 crc kubenswrapper[4805]: I1128 15:42:11.061830 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" containerID="cri-o://9401a6879e1fcc43d0b27657309da77bf635d82861975cc53606a87548ec258e" gracePeriod=600 Nov 28 15:42:11 crc kubenswrapper[4805]: I1128 15:42:11.218850 4805 generic.go:334] "Generic (PLEG): container finished" podID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerID="9401a6879e1fcc43d0b27657309da77bf635d82861975cc53606a87548ec258e" exitCode=0 Nov 28 15:42:11 crc kubenswrapper[4805]: I1128 15:42:11.225019 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerDied","Data":"9401a6879e1fcc43d0b27657309da77bf635d82861975cc53606a87548ec258e"} Nov 28 15:42:11 crc kubenswrapper[4805]: I1128 15:42:11.225104 4805 scope.go:117] "RemoveContainer" containerID="1059ec0f76219a39add1911f16614ef92240c8775a67469a3bc7df93d8da629e" Nov 28 15:42:12 crc kubenswrapper[4805]: I1128 15:42:12.228664 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"ada3fc8f488f7799cae44933973db15a3177119e0d7d2d4556df785008a6700d"} Nov 28 15:42:12 crc kubenswrapper[4805]: I1128 15:42:12.248295 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" podStartSLOduration=5.8323396800000005 podStartE2EDuration="11.248268292s" podCreationTimestamp="2025-11-28 15:42:01 +0000 UTC" firstStartedPulling="2025-11-28 15:42:01.61616954 +0000 UTC m=+948.665960851" lastFinishedPulling="2025-11-28 15:42:07.032098142 +0000 UTC m=+954.081889463" observedRunningTime="2025-11-28 15:42:07.217080932 +0000 UTC m=+954.266872253" watchObservedRunningTime="2025-11-28 15:42:12.248268292 +0000 UTC m=+959.298059623" Nov 28 15:42:21 crc kubenswrapper[4805]: I1128 15:42:21.374394 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-848c47db98-7l2qm" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.176302 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.178137 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.181265 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.181599 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-bvwd2" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.182654 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.183866 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-4xtt9" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.193420 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-78jxp"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.194666 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.197257 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.200074 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-9lzjm" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.224600 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.225491 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.229863 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-stcdm" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.249035 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.261514 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-78jxp"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.274093 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.290415 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.291562 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.294678 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9zsg\" (UniqueName: \"kubernetes.io/projected/6d7afdad-6a62-4c69-b453-dcf5b917a568-kube-api-access-g9zsg\") pod \"cinder-operator-controller-manager-6b7f75547b-m6cln\" (UID: \"6d7afdad-6a62-4c69-b453-dcf5b917a568\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.294735 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkvdn\" (UniqueName: \"kubernetes.io/projected/697a2a64-1776-4bb4-95f4-901e1ae2bbc2-kube-api-access-gkvdn\") pod \"designate-operator-controller-manager-955677c94-78jxp\" (UID: \"697a2a64-1776-4bb4-95f4-901e1ae2bbc2\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.294758 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mssgs\" (UniqueName: \"kubernetes.io/projected/8ff9654c-7de4-4cc2-af76-863ac6cb8694-kube-api-access-mssgs\") pod \"glance-operator-controller-manager-589cbd6b5b-p9lc4\" (UID: \"8ff9654c-7de4-4cc2-af76-863ac6cb8694\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.294786 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpdc8\" (UniqueName: \"kubernetes.io/projected/446d7f87-a4e9-4a87-9091-926345489a0d-kube-api-access-jpdc8\") pod \"barbican-operator-controller-manager-7b64f4fb85-lh9jk\" (UID: \"446d7f87-a4e9-4a87-9091-926345489a0d\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.295463 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-tzndl" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.301156 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.302459 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.318797 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.329342 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mx74b" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.349372 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.349560 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.366220 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gfbgt" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.366405 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.368984 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.397117 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d8pz\" (UniqueName: \"kubernetes.io/projected/d786fb2a-908c-42bc-8889-f85d48a824c3-kube-api-access-9d8pz\") pod \"heat-operator-controller-manager-5b77f656f-jl2bq\" (UID: \"d786fb2a-908c-42bc-8889-f85d48a824c3\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.397408 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9zsg\" (UniqueName: \"kubernetes.io/projected/6d7afdad-6a62-4c69-b453-dcf5b917a568-kube-api-access-g9zsg\") pod \"cinder-operator-controller-manager-6b7f75547b-m6cln\" (UID: \"6d7afdad-6a62-4c69-b453-dcf5b917a568\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.397520 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.397640 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkvdn\" (UniqueName: \"kubernetes.io/projected/697a2a64-1776-4bb4-95f4-901e1ae2bbc2-kube-api-access-gkvdn\") pod \"designate-operator-controller-manager-955677c94-78jxp\" (UID: \"697a2a64-1776-4bb4-95f4-901e1ae2bbc2\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.397747 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mssgs\" (UniqueName: \"kubernetes.io/projected/8ff9654c-7de4-4cc2-af76-863ac6cb8694-kube-api-access-mssgs\") pod \"glance-operator-controller-manager-589cbd6b5b-p9lc4\" (UID: \"8ff9654c-7de4-4cc2-af76-863ac6cb8694\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.397869 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpdc8\" (UniqueName: \"kubernetes.io/projected/446d7f87-a4e9-4a87-9091-926345489a0d-kube-api-access-jpdc8\") pod \"barbican-operator-controller-manager-7b64f4fb85-lh9jk\" (UID: \"446d7f87-a4e9-4a87-9091-926345489a0d\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.397983 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ncfz\" (UniqueName: \"kubernetes.io/projected/4a95d51b-3298-4575-84d3-5b1f1174edbe-kube-api-access-9ncfz\") pod \"horizon-operator-controller-manager-5d494799bf-jn44r\" (UID: \"4a95d51b-3298-4575-84d3-5b1f1174edbe\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.398112 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wtfm\" (UniqueName: \"kubernetes.io/projected/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-kube-api-access-2wtfm\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.433712 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9zsg\" (UniqueName: \"kubernetes.io/projected/6d7afdad-6a62-4c69-b453-dcf5b917a568-kube-api-access-g9zsg\") pod \"cinder-operator-controller-manager-6b7f75547b-m6cln\" (UID: \"6d7afdad-6a62-4c69-b453-dcf5b917a568\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.433782 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpdc8\" (UniqueName: \"kubernetes.io/projected/446d7f87-a4e9-4a87-9091-926345489a0d-kube-api-access-jpdc8\") pod \"barbican-operator-controller-manager-7b64f4fb85-lh9jk\" (UID: \"446d7f87-a4e9-4a87-9091-926345489a0d\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.441002 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mssgs\" (UniqueName: \"kubernetes.io/projected/8ff9654c-7de4-4cc2-af76-863ac6cb8694-kube-api-access-mssgs\") pod \"glance-operator-controller-manager-589cbd6b5b-p9lc4\" (UID: \"8ff9654c-7de4-4cc2-af76-863ac6cb8694\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.442932 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkvdn\" (UniqueName: \"kubernetes.io/projected/697a2a64-1776-4bb4-95f4-901e1ae2bbc2-kube-api-access-gkvdn\") pod \"designate-operator-controller-manager-955677c94-78jxp\" (UID: \"697a2a64-1776-4bb4-95f4-901e1ae2bbc2\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.448503 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.449429 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.454168 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-jgsk8" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.468487 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.474434 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.475426 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.479713 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-xs28l" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.492409 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.498539 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.498639 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.498928 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ncfz\" (UniqueName: \"kubernetes.io/projected/4a95d51b-3298-4575-84d3-5b1f1174edbe-kube-api-access-9ncfz\") pod \"horizon-operator-controller-manager-5d494799bf-jn44r\" (UID: \"4a95d51b-3298-4575-84d3-5b1f1174edbe\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.498977 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wtfm\" (UniqueName: \"kubernetes.io/projected/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-kube-api-access-2wtfm\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.499016 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnk5b\" (UniqueName: \"kubernetes.io/projected/cd24b067-2061-47eb-8494-5a531f13b111-kube-api-access-tnk5b\") pod \"keystone-operator-controller-manager-7b4567c7cf-5m48r\" (UID: \"cd24b067-2061-47eb-8494-5a531f13b111\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.499053 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d8pz\" (UniqueName: \"kubernetes.io/projected/d786fb2a-908c-42bc-8889-f85d48a824c3-kube-api-access-9d8pz\") pod \"heat-operator-controller-manager-5b77f656f-jl2bq\" (UID: \"d786fb2a-908c-42bc-8889-f85d48a824c3\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.499075 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.512535 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf69b\" (UniqueName: \"kubernetes.io/projected/f0bf9b46-6d36-47b4-bbac-374d79b553fc-kube-api-access-rf69b\") pod \"ironic-operator-controller-manager-67cb4dc6d4-v7f97\" (UID: \"f0bf9b46-6d36-47b4-bbac-374d79b553fc\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.507431 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.512890 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" Nov 28 15:42:39 crc kubenswrapper[4805]: E1128 15:42:39.499491 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:39 crc kubenswrapper[4805]: E1128 15:42:39.513628 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert podName:dfdaba79-b86b-4a03-87e0-ed9f712ea4bb nodeName:}" failed. No retries permitted until 2025-11-28 15:42:40.013609719 +0000 UTC m=+987.063401030 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert") pod "infra-operator-controller-manager-57548d458d-2p9q6" (UID: "dfdaba79-b86b-4a03-87e0-ed9f712ea4bb") : secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.514067 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.522228 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-t4vsh" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.525333 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.525903 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d8pz\" (UniqueName: \"kubernetes.io/projected/d786fb2a-908c-42bc-8889-f85d48a824c3-kube-api-access-9d8pz\") pod \"heat-operator-controller-manager-5b77f656f-jl2bq\" (UID: \"d786fb2a-908c-42bc-8889-f85d48a824c3\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.526926 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ncfz\" (UniqueName: \"kubernetes.io/projected/4a95d51b-3298-4575-84d3-5b1f1174edbe-kube-api-access-9ncfz\") pod \"horizon-operator-controller-manager-5d494799bf-jn44r\" (UID: \"4a95d51b-3298-4575-84d3-5b1f1174edbe\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.531048 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.532413 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wtfm\" (UniqueName: \"kubernetes.io/projected/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-kube-api-access-2wtfm\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.532691 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.534680 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-dlfn4" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.540717 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.547930 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.554430 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.556452 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.559632 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-r45z8" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.568634 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.573451 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.574549 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.576005 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-cfh5k" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.591902 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.605377 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.606940 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.607272 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.609448 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-48srn" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.612911 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.613510 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42tkr\" (UniqueName: \"kubernetes.io/projected/49986953-d4ca-4e5f-ac36-f91d86098ff5-kube-api-access-42tkr\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-skdkr\" (UID: \"49986953-d4ca-4e5f-ac36-f91d86098ff5\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.614011 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnk5b\" (UniqueName: \"kubernetes.io/projected/cd24b067-2061-47eb-8494-5a531f13b111-kube-api-access-tnk5b\") pod \"keystone-operator-controller-manager-7b4567c7cf-5m48r\" (UID: \"cd24b067-2061-47eb-8494-5a531f13b111\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.614097 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5dn2\" (UniqueName: \"kubernetes.io/projected/75b25042-ef68-47da-873b-6750083ecc7e-kube-api-access-j5dn2\") pod \"neutron-operator-controller-manager-6fdcddb789-lkqjk\" (UID: \"75b25042-ef68-47da-873b-6750083ecc7e\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.614149 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxtwr\" (UniqueName: \"kubernetes.io/projected/0dc0e269-4d2b-4e4b-afcc-0547ac908a26-kube-api-access-gxtwr\") pod \"nova-operator-controller-manager-79556f57fc-2wbp7\" (UID: \"0dc0e269-4d2b-4e4b-afcc-0547ac908a26\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.614183 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mdsm\" (UniqueName: \"kubernetes.io/projected/3bad1d97-f824-4616-a08d-9e00fd2c5201-kube-api-access-2mdsm\") pod \"manila-operator-controller-manager-5d499bf58b-bqmpw\" (UID: \"3bad1d97-f824-4616-a08d-9e00fd2c5201\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.614213 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf69b\" (UniqueName: \"kubernetes.io/projected/f0bf9b46-6d36-47b4-bbac-374d79b553fc-kube-api-access-rf69b\") pod \"ironic-operator-controller-manager-67cb4dc6d4-v7f97\" (UID: \"f0bf9b46-6d36-47b4-bbac-374d79b553fc\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.634443 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf69b\" (UniqueName: \"kubernetes.io/projected/f0bf9b46-6d36-47b4-bbac-374d79b553fc-kube-api-access-rf69b\") pod \"ironic-operator-controller-manager-67cb4dc6d4-v7f97\" (UID: \"f0bf9b46-6d36-47b4-bbac-374d79b553fc\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.637245 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnk5b\" (UniqueName: \"kubernetes.io/projected/cd24b067-2061-47eb-8494-5a531f13b111-kube-api-access-tnk5b\") pod \"keystone-operator-controller-manager-7b4567c7cf-5m48r\" (UID: \"cd24b067-2061-47eb-8494-5a531f13b111\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.639474 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.651329 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.653188 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.656530 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-77xw5" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.656670 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.664021 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.666093 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.673944 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.674880 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-4ld5j" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.675085 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.677030 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.679291 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-2fwl4" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.690847 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.699178 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.724425 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.729867 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5dn2\" (UniqueName: \"kubernetes.io/projected/75b25042-ef68-47da-873b-6750083ecc7e-kube-api-access-j5dn2\") pod \"neutron-operator-controller-manager-6fdcddb789-lkqjk\" (UID: \"75b25042-ef68-47da-873b-6750083ecc7e\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.730024 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxtwr\" (UniqueName: \"kubernetes.io/projected/0dc0e269-4d2b-4e4b-afcc-0547ac908a26-kube-api-access-gxtwr\") pod \"nova-operator-controller-manager-79556f57fc-2wbp7\" (UID: \"0dc0e269-4d2b-4e4b-afcc-0547ac908a26\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.730107 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mdsm\" (UniqueName: \"kubernetes.io/projected/3bad1d97-f824-4616-a08d-9e00fd2c5201-kube-api-access-2mdsm\") pod \"manila-operator-controller-manager-5d499bf58b-bqmpw\" (UID: \"3bad1d97-f824-4616-a08d-9e00fd2c5201\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.730236 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.730395 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf9l2\" (UniqueName: \"kubernetes.io/projected/5cb86467-0536-46c2-a399-53bf3ea553b3-kube-api-access-lf9l2\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.730537 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7x2t\" (UniqueName: \"kubernetes.io/projected/2dac7bca-c6db-40a9-ac64-b46c7722798b-kube-api-access-h7x2t\") pod \"octavia-operator-controller-manager-64cdc6ff96-c6lqn\" (UID: \"2dac7bca-c6db-40a9-ac64-b46c7722798b\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.730619 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42tkr\" (UniqueName: \"kubernetes.io/projected/49986953-d4ca-4e5f-ac36-f91d86098ff5-kube-api-access-42tkr\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-skdkr\" (UID: \"49986953-d4ca-4e5f-ac36-f91d86098ff5\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.730707 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcj25\" (UniqueName: \"kubernetes.io/projected/8e874a32-dff0-48ad-b43b-a24c364e5e8d-kube-api-access-hcj25\") pod \"ovn-operator-controller-manager-56897c768d-9q29v\" (UID: \"8e874a32-dff0-48ad-b43b-a24c364e5e8d\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.730788 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68bxl\" (UniqueName: \"kubernetes.io/projected/abe59027-7c48-4b8e-8255-067032ba364d-kube-api-access-68bxl\") pod \"placement-operator-controller-manager-57988cc5b5-hnhwx\" (UID: \"abe59027-7c48-4b8e-8255-067032ba364d\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.734874 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.736004 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.741163 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-ljvmh" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.756244 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.789970 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5dn2\" (UniqueName: \"kubernetes.io/projected/75b25042-ef68-47da-873b-6750083ecc7e-kube-api-access-j5dn2\") pod \"neutron-operator-controller-manager-6fdcddb789-lkqjk\" (UID: \"75b25042-ef68-47da-873b-6750083ecc7e\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.798840 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.799947 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.802922 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.803454 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mdsm\" (UniqueName: \"kubernetes.io/projected/3bad1d97-f824-4616-a08d-9e00fd2c5201-kube-api-access-2mdsm\") pod \"manila-operator-controller-manager-5d499bf58b-bqmpw\" (UID: \"3bad1d97-f824-4616-a08d-9e00fd2c5201\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.803863 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxtwr\" (UniqueName: \"kubernetes.io/projected/0dc0e269-4d2b-4e4b-afcc-0547ac908a26-kube-api-access-gxtwr\") pod \"nova-operator-controller-manager-79556f57fc-2wbp7\" (UID: \"0dc0e269-4d2b-4e4b-afcc-0547ac908a26\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.805542 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-5ndth" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.808028 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42tkr\" (UniqueName: \"kubernetes.io/projected/49986953-d4ca-4e5f-ac36-f91d86098ff5-kube-api-access-42tkr\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-skdkr\" (UID: \"49986953-d4ca-4e5f-ac36-f91d86098ff5\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.817931 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.830844 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.831949 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.834412 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swl86\" (UniqueName: \"kubernetes.io/projected/5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0-kube-api-access-swl86\") pod \"swift-operator-controller-manager-d77b94747-mg6cr\" (UID: \"5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.834458 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7x2t\" (UniqueName: \"kubernetes.io/projected/2dac7bca-c6db-40a9-ac64-b46c7722798b-kube-api-access-h7x2t\") pod \"octavia-operator-controller-manager-64cdc6ff96-c6lqn\" (UID: \"2dac7bca-c6db-40a9-ac64-b46c7722798b\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.834486 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcj25\" (UniqueName: \"kubernetes.io/projected/8e874a32-dff0-48ad-b43b-a24c364e5e8d-kube-api-access-hcj25\") pod \"ovn-operator-controller-manager-56897c768d-9q29v\" (UID: \"8e874a32-dff0-48ad-b43b-a24c364e5e8d\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.834517 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68bxl\" (UniqueName: \"kubernetes.io/projected/abe59027-7c48-4b8e-8255-067032ba364d-kube-api-access-68bxl\") pod \"placement-operator-controller-manager-57988cc5b5-hnhwx\" (UID: \"abe59027-7c48-4b8e-8255-067032ba364d\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.834547 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9vwj\" (UniqueName: \"kubernetes.io/projected/01f64db7-5c9e-41ea-a595-75dd81511543-kube-api-access-z9vwj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-2jrp8\" (UID: \"01f64db7-5c9e-41ea-a595-75dd81511543\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.834602 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.834619 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf9l2\" (UniqueName: \"kubernetes.io/projected/5cb86467-0536-46c2-a399-53bf3ea553b3-kube-api-access-lf9l2\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:39 crc kubenswrapper[4805]: E1128 15:42:39.835220 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:39 crc kubenswrapper[4805]: E1128 15:42:39.835265 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert podName:5cb86467-0536-46c2-a399-53bf3ea553b3 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:40.335252206 +0000 UTC m=+987.385043517 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" (UID: "5cb86467-0536-46c2-a399-53bf3ea553b3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.841485 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-2nj8f" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.846835 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.875872 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf9l2\" (UniqueName: \"kubernetes.io/projected/5cb86467-0536-46c2-a399-53bf3ea553b3-kube-api-access-lf9l2\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.880161 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7x2t\" (UniqueName: \"kubernetes.io/projected/2dac7bca-c6db-40a9-ac64-b46c7722798b-kube-api-access-h7x2t\") pod \"octavia-operator-controller-manager-64cdc6ff96-c6lqn\" (UID: \"2dac7bca-c6db-40a9-ac64-b46c7722798b\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.885694 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68bxl\" (UniqueName: \"kubernetes.io/projected/abe59027-7c48-4b8e-8255-067032ba364d-kube-api-access-68bxl\") pod \"placement-operator-controller-manager-57988cc5b5-hnhwx\" (UID: \"abe59027-7c48-4b8e-8255-067032ba364d\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.897722 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcj25\" (UniqueName: \"kubernetes.io/projected/8e874a32-dff0-48ad-b43b-a24c364e5e8d-kube-api-access-hcj25\") pod \"ovn-operator-controller-manager-56897c768d-9q29v\" (UID: \"8e874a32-dff0-48ad-b43b-a24c364e5e8d\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.910921 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.929744 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.936961 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9vwj\" (UniqueName: \"kubernetes.io/projected/01f64db7-5c9e-41ea-a595-75dd81511543-kube-api-access-z9vwj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-2jrp8\" (UID: \"01f64db7-5c9e-41ea-a595-75dd81511543\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.937097 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-456tp\" (UniqueName: \"kubernetes.io/projected/dc3a6e21-e0ca-40d0-b140-21215085f97a-kube-api-access-456tp\") pod \"test-operator-controller-manager-5cd6c7f4c8-pmlh7\" (UID: \"dc3a6e21-e0ca-40d0-b140-21215085f97a\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.937133 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swl86\" (UniqueName: \"kubernetes.io/projected/5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0-kube-api-access-swl86\") pod \"swift-operator-controller-manager-d77b94747-mg6cr\" (UID: \"5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.946627 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.947719 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.952181 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.958095 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.958453 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-ln2k4" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.966859 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.989454 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n"] Nov 28 15:42:39 crc kubenswrapper[4805]: I1128 15:42:39.990537 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.001462 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xww5b" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.001558 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.001465 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.007344 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.017401 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swl86\" (UniqueName: \"kubernetes.io/projected/5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0-kube-api-access-swl86\") pod \"swift-operator-controller-manager-d77b94747-mg6cr\" (UID: \"5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.017833 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.018745 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9vwj\" (UniqueName: \"kubernetes.io/projected/01f64db7-5c9e-41ea-a595-75dd81511543-kube-api-access-z9vwj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-2jrp8\" (UID: \"01f64db7-5c9e-41ea-a595-75dd81511543\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.038715 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.039033 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.039150 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qhh4\" (UniqueName: \"kubernetes.io/projected/519f14b3-c5a1-4bed-8edc-92a84bf0a70d-kube-api-access-2qhh4\") pod \"watcher-operator-controller-manager-656dcb59d4-2pnsl\" (UID: \"519f14b3-c5a1-4bed-8edc-92a84bf0a70d\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.039586 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7gsf\" (UniqueName: \"kubernetes.io/projected/e8c06141-6007-491e-bdd0-6a7654607554-kube-api-access-f7gsf\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.039706 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-456tp\" (UniqueName: \"kubernetes.io/projected/dc3a6e21-e0ca-40d0-b140-21215085f97a-kube-api-access-456tp\") pod \"test-operator-controller-manager-5cd6c7f4c8-pmlh7\" (UID: \"dc3a6e21-e0ca-40d0-b140-21215085f97a\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.039789 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.039973 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.040070 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert podName:dfdaba79-b86b-4a03-87e0-ed9f712ea4bb nodeName:}" failed. No retries permitted until 2025-11-28 15:42:41.040056627 +0000 UTC m=+988.089847938 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert") pod "infra-operator-controller-manager-57548d458d-2p9q6" (UID: "dfdaba79-b86b-4a03-87e0-ed9f712ea4bb") : secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.043155 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.065588 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.107343 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.117340 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-456tp\" (UniqueName: \"kubernetes.io/projected/dc3a6e21-e0ca-40d0-b140-21215085f97a-kube-api-access-456tp\") pod \"test-operator-controller-manager-5cd6c7f4c8-pmlh7\" (UID: \"dc3a6e21-e0ca-40d0-b140-21215085f97a\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.132252 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.137933 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.140616 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.140729 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.140794 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qhh4\" (UniqueName: \"kubernetes.io/projected/519f14b3-c5a1-4bed-8edc-92a84bf0a70d-kube-api-access-2qhh4\") pod \"watcher-operator-controller-manager-656dcb59d4-2pnsl\" (UID: \"519f14b3-c5a1-4bed-8edc-92a84bf0a70d\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.140818 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7gsf\" (UniqueName: \"kubernetes.io/projected/e8c06141-6007-491e-bdd0-6a7654607554-kube-api-access-f7gsf\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.141252 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.141296 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:40.641283455 +0000 UTC m=+987.691074766 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "webhook-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.141695 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.141772 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:40.641746547 +0000 UTC m=+987.691537938 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "metrics-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.180251 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.181394 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.183882 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.190915 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-hqrf9" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.191985 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qhh4\" (UniqueName: \"kubernetes.io/projected/519f14b3-c5a1-4bed-8edc-92a84bf0a70d-kube-api-access-2qhh4\") pod \"watcher-operator-controller-manager-656dcb59d4-2pnsl\" (UID: \"519f14b3-c5a1-4bed-8edc-92a84bf0a70d\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.194416 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.196303 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7gsf\" (UniqueName: \"kubernetes.io/projected/e8c06141-6007-491e-bdd0-6a7654607554-kube-api-access-f7gsf\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.244909 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q96nf\" (UniqueName: \"kubernetes.io/projected/25d77f9e-dbcf-42b8-8e8e-28d3a500791d-kube-api-access-q96nf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-m9thn\" (UID: \"25d77f9e-dbcf-42b8-8e8e-28d3a500791d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.349316 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q96nf\" (UniqueName: \"kubernetes.io/projected/25d77f9e-dbcf-42b8-8e8e-28d3a500791d-kube-api-access-q96nf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-m9thn\" (UID: \"25d77f9e-dbcf-42b8-8e8e-28d3a500791d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.349436 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.349561 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.349604 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert podName:5cb86467-0536-46c2-a399-53bf3ea553b3 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:41.349591531 +0000 UTC m=+988.399382842 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" (UID: "5cb86467-0536-46c2-a399-53bf3ea553b3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.371229 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q96nf\" (UniqueName: \"kubernetes.io/projected/25d77f9e-dbcf-42b8-8e8e-28d3a500791d-kube-api-access-q96nf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-m9thn\" (UID: \"25d77f9e-dbcf-42b8-8e8e-28d3a500791d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.471003 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.479698 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.516696 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-78jxp"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.527398 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.534602 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.653849 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.653956 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.654114 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.654161 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:41.654142818 +0000 UTC m=+988.703934129 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "webhook-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.654430 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: E1128 15:42:40.654515 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:41.654494858 +0000 UTC m=+988.704286169 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "metrics-server-cert" not found Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.914623 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.942881 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.968096 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.979855 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r"] Nov 28 15:42:40 crc kubenswrapper[4805]: I1128 15:42:40.986070 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr"] Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.004062 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk"] Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.016067 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7"] Nov 28 15:42:41 crc kubenswrapper[4805]: W1128 15:42:41.018837 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dc0e269_4d2b_4e4b_afcc_0547ac908a26.slice/crio-00c63e454de9b39605d0d8e29c2f12b8008ffdaacfbccc6e22d07754502756a2 WatchSource:0}: Error finding container 00c63e454de9b39605d0d8e29c2f12b8008ffdaacfbccc6e22d07754502756a2: Status 404 returned error can't find the container with id 00c63e454de9b39605d0d8e29c2f12b8008ffdaacfbccc6e22d07754502756a2 Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.058767 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.058975 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.059054 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert podName:dfdaba79-b86b-4a03-87e0-ed9f712ea4bb nodeName:}" failed. No retries permitted until 2025-11-28 15:42:43.059034408 +0000 UTC m=+990.108825709 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert") pod "infra-operator-controller-manager-57548d458d-2p9q6" (UID: "dfdaba79-b86b-4a03-87e0-ed9f712ea4bb") : secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.183349 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx"] Nov 28 15:42:41 crc kubenswrapper[4805]: W1128 15:42:41.189466 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabe59027_7c48_4b8e_8255_067032ba364d.slice/crio-4a11af69696c72c9ee1467cf58e4cc7c61865ea81a7db9b8662ea1862b89d874 WatchSource:0}: Error finding container 4a11af69696c72c9ee1467cf58e4cc7c61865ea81a7db9b8662ea1862b89d874: Status 404 returned error can't find the container with id 4a11af69696c72c9ee1467cf58e4cc7c61865ea81a7db9b8662ea1862b89d874 Nov 28 15:42:41 crc kubenswrapper[4805]: W1128 15:42:41.194192 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd24b067_2061_47eb_8494_5a531f13b111.slice/crio-7c08ea41fa759258251f2ac0bd010117b4ac2f15686d79374f8736111fd7c1e9 WatchSource:0}: Error finding container 7c08ea41fa759258251f2ac0bd010117b4ac2f15686d79374f8736111fd7c1e9: Status 404 returned error can't find the container with id 7c08ea41fa759258251f2ac0bd010117b4ac2f15686d79374f8736111fd7c1e9 Nov 28 15:42:41 crc kubenswrapper[4805]: W1128 15:42:41.194811 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a4bcea2_6232_49b3_89c9_09fbcb1ae5a0.slice/crio-cec644ac85535ef8ffbcb9cbb0142420718021307ab9714138776fc42d292e46 WatchSource:0}: Error finding container cec644ac85535ef8ffbcb9cbb0142420718021307ab9714138776fc42d292e46: Status 404 returned error can't find the container with id cec644ac85535ef8ffbcb9cbb0142420718021307ab9714138776fc42d292e46 Nov 28 15:42:41 crc kubenswrapper[4805]: W1128 15:42:41.196735 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01f64db7_5c9e_41ea_a595_75dd81511543.slice/crio-9cb01804b266e541c361cc4acfb6dca60a864ace4e4dd3d2457f08a6787f30ce WatchSource:0}: Error finding container 9cb01804b266e541c361cc4acfb6dca60a864ace4e4dd3d2457f08a6787f30ce: Status 404 returned error can't find the container with id 9cb01804b266e541c361cc4acfb6dca60a864ace4e4dd3d2457f08a6787f30ce Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.198755 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tnk5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7b4567c7cf-5m48r_openstack-operators(cd24b067-2061-47eb-8494-5a531f13b111): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.200412 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r"] Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.205324 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr"] Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.206550 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tnk5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7b4567c7cf-5m48r_openstack-operators(cd24b067-2061-47eb-8494-5a531f13b111): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.207762 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" podUID="cd24b067-2061-47eb-8494-5a531f13b111" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.216086 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8"] Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.363311 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.363500 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.363557 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert podName:5cb86467-0536-46c2-a399-53bf3ea553b3 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:43.363543595 +0000 UTC m=+990.413334896 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" (UID: "5cb86467-0536-46c2-a399-53bf3ea553b3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.370712 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn"] Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.378298 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw"] Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.391165 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:89910bc3ecceb7590d3207ac294eb7354de358cf39ef03c72323b26c598e50e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2mdsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5d499bf58b-bqmpw_openstack-operators(3bad1d97-f824-4616-a08d-9e00fd2c5201): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.394259 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2mdsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5d499bf58b-bqmpw_openstack-operators(3bad1d97-f824-4616-a08d-9e00fd2c5201): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.394501 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn"] Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.395771 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" podUID="3bad1d97-f824-4616-a08d-9e00fd2c5201" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.401945 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7"] Nov 28 15:42:41 crc kubenswrapper[4805]: W1128 15:42:41.405653 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc3a6e21_e0ca_40d0_b140_21215085f97a.slice/crio-ede5c2b91c622709d1f08d0af6df114865c7621c8625f42ada454b4cd04a135f WatchSource:0}: Error finding container ede5c2b91c622709d1f08d0af6df114865c7621c8625f42ada454b4cd04a135f: Status 404 returned error can't find the container with id ede5c2b91c622709d1f08d0af6df114865c7621c8625f42ada454b4cd04a135f Nov 28 15:42:41 crc kubenswrapper[4805]: W1128 15:42:41.407689 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25d77f9e_dbcf_42b8_8e8e_28d3a500791d.slice/crio-157c7436b9e0d426b33eb658777cd2b86f52573433e4e36386f451f9e83c659a WatchSource:0}: Error finding container 157c7436b9e0d426b33eb658777cd2b86f52573433e4e36386f451f9e83c659a: Status 404 returned error can't find the container with id 157c7436b9e0d426b33eb658777cd2b86f52573433e4e36386f451f9e83c659a Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.408000 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v"] Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.410943 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-456tp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-pmlh7_openstack-operators(dc3a6e21-e0ca-40d0-b140-21215085f97a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.411040 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q96nf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-m9thn_openstack-operators(25d77f9e-dbcf-42b8-8e8e-28d3a500791d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.412099 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hcj25,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-9q29v_openstack-operators(8e874a32-dff0-48ad-b43b-a24c364e5e8d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.412221 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" podUID="25d77f9e-dbcf-42b8-8e8e-28d3a500791d" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.412833 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-456tp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-pmlh7_openstack-operators(dc3a6e21-e0ca-40d0-b140-21215085f97a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.415026 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" podUID="dc3a6e21-e0ca-40d0-b140-21215085f97a" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.415059 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hcj25,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-9q29v_openstack-operators(8e874a32-dff0-48ad-b43b-a24c364e5e8d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.416347 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" podUID="8e874a32-dff0-48ad-b43b-a24c364e5e8d" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.420855 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl"] Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.430525 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2qhh4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-2pnsl_openstack-operators(519f14b3-c5a1-4bed-8edc-92a84bf0a70d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.432636 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2qhh4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-2pnsl_openstack-operators(519f14b3-c5a1-4bed-8edc-92a84bf0a70d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.433991 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" podUID="519f14b3-c5a1-4bed-8edc-92a84bf0a70d" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.444569 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" event={"ID":"dc3a6e21-e0ca-40d0-b140-21215085f97a","Type":"ContainerStarted","Data":"ede5c2b91c622709d1f08d0af6df114865c7621c8625f42ada454b4cd04a135f"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.445644 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" event={"ID":"4a95d51b-3298-4575-84d3-5b1f1174edbe","Type":"ContainerStarted","Data":"df48b40d4d39e0a9b57c3bf0f5a30c0435553fc9eb8731c3f13dc4b8456599f6"} Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.446220 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" podUID="dc3a6e21-e0ca-40d0-b140-21215085f97a" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.447929 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" event={"ID":"f0bf9b46-6d36-47b4-bbac-374d79b553fc","Type":"ContainerStarted","Data":"36800364be7a0db5045cdf7ab9fa84750b5edc4a475b8f1097eed9abd69c5dff"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.453198 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" event={"ID":"cd24b067-2061-47eb-8494-5a531f13b111","Type":"ContainerStarted","Data":"7c08ea41fa759258251f2ac0bd010117b4ac2f15686d79374f8736111fd7c1e9"} Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.456311 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" podUID="cd24b067-2061-47eb-8494-5a531f13b111" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.457914 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" event={"ID":"01f64db7-5c9e-41ea-a595-75dd81511543","Type":"ContainerStarted","Data":"9cb01804b266e541c361cc4acfb6dca60a864ace4e4dd3d2457f08a6787f30ce"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.463291 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" event={"ID":"446d7f87-a4e9-4a87-9091-926345489a0d","Type":"ContainerStarted","Data":"cdb92f5d22978a1d46c3512ccb12043bccc0d1ca6e3bfd9f183f854c6a20a5a6"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.471571 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" event={"ID":"697a2a64-1776-4bb4-95f4-901e1ae2bbc2","Type":"ContainerStarted","Data":"2e36833e4f517605e36aabd6e04134555aeae39f361e979ba5dccfb4552fd506"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.474579 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" event={"ID":"2dac7bca-c6db-40a9-ac64-b46c7722798b","Type":"ContainerStarted","Data":"ce1f74f530566c5afe94e60b0e941e00012a4deaaf55ddc4a9f517398d82aadd"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.475856 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" event={"ID":"5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0","Type":"ContainerStarted","Data":"cec644ac85535ef8ffbcb9cbb0142420718021307ab9714138776fc42d292e46"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.476757 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" event={"ID":"d786fb2a-908c-42bc-8889-f85d48a824c3","Type":"ContainerStarted","Data":"9928af6788632fa87d491e01cc1411a8048fb0b3faeae0326e03eb0206267c08"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.478794 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" event={"ID":"49986953-d4ca-4e5f-ac36-f91d86098ff5","Type":"ContainerStarted","Data":"1bcdb1cb1b1e0fe5d683da03b78e78f58e42eda3646527426c952320c9af905a"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.479990 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" event={"ID":"6d7afdad-6a62-4c69-b453-dcf5b917a568","Type":"ContainerStarted","Data":"04fc68a722837cfc5ec3410e31f90b0e4075be15c5736183120cdb1899cea3b0"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.484109 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" event={"ID":"abe59027-7c48-4b8e-8255-067032ba364d","Type":"ContainerStarted","Data":"4a11af69696c72c9ee1467cf58e4cc7c61865ea81a7db9b8662ea1862b89d874"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.485508 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" event={"ID":"25d77f9e-dbcf-42b8-8e8e-28d3a500791d","Type":"ContainerStarted","Data":"157c7436b9e0d426b33eb658777cd2b86f52573433e4e36386f451f9e83c659a"} Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.487806 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" podUID="25d77f9e-dbcf-42b8-8e8e-28d3a500791d" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.497783 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" event={"ID":"8ff9654c-7de4-4cc2-af76-863ac6cb8694","Type":"ContainerStarted","Data":"3b39632977bfd5da2f05e8f9b324b1bee0300a38a106b20df717f8994137d3d0"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.508368 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" event={"ID":"3bad1d97-f824-4616-a08d-9e00fd2c5201","Type":"ContainerStarted","Data":"1f30e4d66e516fe9d891470924884c40f9bfa4de90de9b6bd695057b3df26fe4"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.511665 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" event={"ID":"8e874a32-dff0-48ad-b43b-a24c364e5e8d","Type":"ContainerStarted","Data":"6db7d0e0ea212edc0e7fdff1ded00e0af9f8f19c78803d11d0c4098279563d86"} Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.512680 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:89910bc3ecceb7590d3207ac294eb7354de358cf39ef03c72323b26c598e50e6\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" podUID="3bad1d97-f824-4616-a08d-9e00fd2c5201" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.515876 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" podUID="8e874a32-dff0-48ad-b43b-a24c364e5e8d" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.516494 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" event={"ID":"0dc0e269-4d2b-4e4b-afcc-0547ac908a26","Type":"ContainerStarted","Data":"00c63e454de9b39605d0d8e29c2f12b8008ffdaacfbccc6e22d07754502756a2"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.517768 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" event={"ID":"75b25042-ef68-47da-873b-6750083ecc7e","Type":"ContainerStarted","Data":"ad10421b463e3ff146c393befa5c14e67485786830f1da3709d402e944da4c70"} Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.519441 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" event={"ID":"519f14b3-c5a1-4bed-8edc-92a84bf0a70d","Type":"ContainerStarted","Data":"35d6d2c82ed598326fa96fa0976868de1de266e15e7595b97b20437493a40ff0"} Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.522036 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" podUID="519f14b3-c5a1-4bed-8edc-92a84bf0a70d" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.666287 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:41 crc kubenswrapper[4805]: I1128 15:42:41.666491 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.666844 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.666972 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:43.666914739 +0000 UTC m=+990.716706080 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "metrics-server-cert" not found Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.667651 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 15:42:41 crc kubenswrapper[4805]: E1128 15:42:41.667736 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:43.667711882 +0000 UTC m=+990.717503233 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "webhook-server-cert" not found Nov 28 15:42:42 crc kubenswrapper[4805]: E1128 15:42:42.529168 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" podUID="25d77f9e-dbcf-42b8-8e8e-28d3a500791d" Nov 28 15:42:42 crc kubenswrapper[4805]: E1128 15:42:42.529775 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:89910bc3ecceb7590d3207ac294eb7354de358cf39ef03c72323b26c598e50e6\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" podUID="3bad1d97-f824-4616-a08d-9e00fd2c5201" Nov 28 15:42:42 crc kubenswrapper[4805]: E1128 15:42:42.529740 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" podUID="519f14b3-c5a1-4bed-8edc-92a84bf0a70d" Nov 28 15:42:42 crc kubenswrapper[4805]: E1128 15:42:42.529937 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" podUID="8e874a32-dff0-48ad-b43b-a24c364e5e8d" Nov 28 15:42:42 crc kubenswrapper[4805]: E1128 15:42:42.529971 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" podUID="dc3a6e21-e0ca-40d0-b140-21215085f97a" Nov 28 15:42:42 crc kubenswrapper[4805]: E1128 15:42:42.530604 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" podUID="cd24b067-2061-47eb-8494-5a531f13b111" Nov 28 15:42:43 crc kubenswrapper[4805]: I1128 15:42:43.094218 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:43 crc kubenswrapper[4805]: E1128 15:42:43.094367 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:43 crc kubenswrapper[4805]: E1128 15:42:43.094426 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert podName:dfdaba79-b86b-4a03-87e0-ed9f712ea4bb nodeName:}" failed. No retries permitted until 2025-11-28 15:42:47.094410551 +0000 UTC m=+994.144201862 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert") pod "infra-operator-controller-manager-57548d458d-2p9q6" (UID: "dfdaba79-b86b-4a03-87e0-ed9f712ea4bb") : secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:43 crc kubenswrapper[4805]: I1128 15:42:43.398704 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:43 crc kubenswrapper[4805]: E1128 15:42:43.398864 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:43 crc kubenswrapper[4805]: E1128 15:42:43.398939 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert podName:5cb86467-0536-46c2-a399-53bf3ea553b3 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:47.398920598 +0000 UTC m=+994.448711909 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" (UID: "5cb86467-0536-46c2-a399-53bf3ea553b3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:43 crc kubenswrapper[4805]: I1128 15:42:43.703452 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:43 crc kubenswrapper[4805]: I1128 15:42:43.703563 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:43 crc kubenswrapper[4805]: E1128 15:42:43.703706 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 15:42:43 crc kubenswrapper[4805]: E1128 15:42:43.703753 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:47.703738791 +0000 UTC m=+994.753530102 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "webhook-server-cert" not found Nov 28 15:42:43 crc kubenswrapper[4805]: E1128 15:42:43.704084 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 15:42:43 crc kubenswrapper[4805]: E1128 15:42:43.704110 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:47.704101001 +0000 UTC m=+994.753892312 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "metrics-server-cert" not found Nov 28 15:42:47 crc kubenswrapper[4805]: I1128 15:42:47.152099 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:47 crc kubenswrapper[4805]: E1128 15:42:47.152291 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:47 crc kubenswrapper[4805]: E1128 15:42:47.152656 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert podName:dfdaba79-b86b-4a03-87e0-ed9f712ea4bb nodeName:}" failed. No retries permitted until 2025-11-28 15:42:55.152641221 +0000 UTC m=+1002.202432532 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert") pod "infra-operator-controller-manager-57548d458d-2p9q6" (UID: "dfdaba79-b86b-4a03-87e0-ed9f712ea4bb") : secret "infra-operator-webhook-server-cert" not found Nov 28 15:42:47 crc kubenswrapper[4805]: I1128 15:42:47.458177 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:47 crc kubenswrapper[4805]: E1128 15:42:47.458451 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:47 crc kubenswrapper[4805]: E1128 15:42:47.458496 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert podName:5cb86467-0536-46c2-a399-53bf3ea553b3 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:55.458482845 +0000 UTC m=+1002.508274156 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert") pod "openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" (UID: "5cb86467-0536-46c2-a399-53bf3ea553b3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 15:42:47 crc kubenswrapper[4805]: I1128 15:42:47.761694 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:47 crc kubenswrapper[4805]: I1128 15:42:47.761770 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:47 crc kubenswrapper[4805]: E1128 15:42:47.761889 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 15:42:47 crc kubenswrapper[4805]: E1128 15:42:47.761932 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:55.761918881 +0000 UTC m=+1002.811710192 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "metrics-server-cert" not found Nov 28 15:42:47 crc kubenswrapper[4805]: E1128 15:42:47.762226 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 15:42:47 crc kubenswrapper[4805]: E1128 15:42:47.762255 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:42:55.76224819 +0000 UTC m=+1002.812039501 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "webhook-server-cert" not found Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.386657 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-68bxl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-hnhwx_openstack-operators(abe59027-7c48-4b8e-8255-067032ba364d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.387239 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gkvdn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-955677c94-78jxp_openstack-operators(697a2a64-1776-4bb4-95f4-901e1ae2bbc2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.388433 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" podUID="697a2a64-1776-4bb4-95f4-901e1ae2bbc2" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.388493 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" podUID="abe59027-7c48-4b8e-8255-067032ba364d" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.394184 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rf69b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-67cb4dc6d4-v7f97_openstack-operators(f0bf9b46-6d36-47b4-bbac-374d79b553fc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.396756 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" podUID="f0bf9b46-6d36-47b4-bbac-374d79b553fc" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.404152 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h7x2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-64cdc6ff96-c6lqn_openstack-operators(2dac7bca-c6db-40a9-ac64-b46c7722798b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.405433 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" podUID="2dac7bca-c6db-40a9-ac64-b46c7722798b" Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.604842 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" event={"ID":"697a2a64-1776-4bb4-95f4-901e1ae2bbc2","Type":"ContainerStarted","Data":"4ba908631ddd95808b5ecb4e85dc0e20825cc60464ec2173c28889f4d30070b9"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.604971 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.606480 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" event={"ID":"6d7afdad-6a62-4c69-b453-dcf5b917a568","Type":"ContainerStarted","Data":"445cf12962033ab5c90872a9e3796b1f94888ae7e3776ea5ba92c95a20299fdf"} Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.607308 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" podUID="697a2a64-1776-4bb4-95f4-901e1ae2bbc2" Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.608126 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" event={"ID":"2dac7bca-c6db-40a9-ac64-b46c7722798b","Type":"ContainerStarted","Data":"f1e90edab2af4768839578102f82b394e3624b903f533c32821432e55980831f"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.608273 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.609437 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" podUID="2dac7bca-c6db-40a9-ac64-b46c7722798b" Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.609629 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" event={"ID":"f0bf9b46-6d36-47b4-bbac-374d79b553fc","Type":"ContainerStarted","Data":"49ffbc5ae01c678358bf7df9edf7b09af19d930a594c3980e6705deafeaecb6e"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.609713 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.610573 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" podUID="f0bf9b46-6d36-47b4-bbac-374d79b553fc" Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.611776 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" event={"ID":"0dc0e269-4d2b-4e4b-afcc-0547ac908a26","Type":"ContainerStarted","Data":"e0c61a59150bbf592e62cb0e32762b9aee7e1889e459c7e2b52cbd981f692821"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.613290 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" event={"ID":"4a95d51b-3298-4575-84d3-5b1f1174edbe","Type":"ContainerStarted","Data":"71a1f745ca287f115c52e5154484f5674730abdd161a5ba744a11b55225b12c1"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.615093 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" event={"ID":"01f64db7-5c9e-41ea-a595-75dd81511543","Type":"ContainerStarted","Data":"d3be01277295d9747b786df78a12b6c02b0e70f4b749ffe783ace0feb94c3e41"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.616904 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" event={"ID":"75b25042-ef68-47da-873b-6750083ecc7e","Type":"ContainerStarted","Data":"aec84c2d7a866f138973f7c2d6f105afc54cbcf9e6af9498eb677d2082f068fb"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.618568 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" event={"ID":"49986953-d4ca-4e5f-ac36-f91d86098ff5","Type":"ContainerStarted","Data":"5119ccc6126f2831fafc1f5b2ec340af20093cfbf63be5bea2585524871885f0"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.619957 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" event={"ID":"5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0","Type":"ContainerStarted","Data":"5391c476bd50603e1eba1f45665de7473d1d72088d6e7fb4360552e799c459a1"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.621436 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" event={"ID":"8ff9654c-7de4-4cc2-af76-863ac6cb8694","Type":"ContainerStarted","Data":"3e2fcbed4cb6c0544ea54c6312b86809b08206e63d01389d0af62e79e8f98504"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.623206 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" event={"ID":"d786fb2a-908c-42bc-8889-f85d48a824c3","Type":"ContainerStarted","Data":"419b5110a77b92c8fe7bd2011550db4c920418fd8ea29f7780901542fe44bae6"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.625102 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" event={"ID":"abe59027-7c48-4b8e-8255-067032ba364d","Type":"ContainerStarted","Data":"52b25a8b6dcfaadc6acf18bb5b2d699795efc74059ba61a07a17c405b35dbed4"} Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.625272 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" Nov 28 15:42:53 crc kubenswrapper[4805]: E1128 15:42:53.626395 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" podUID="abe59027-7c48-4b8e-8255-067032ba364d" Nov 28 15:42:53 crc kubenswrapper[4805]: I1128 15:42:53.626763 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" event={"ID":"446d7f87-a4e9-4a87-9091-926345489a0d","Type":"ContainerStarted","Data":"c5c0bd9510a9f837569dde2080d44c4cd065f1ee494fcd77e795441ca3cb5dd6"} Nov 28 15:42:54 crc kubenswrapper[4805]: I1128 15:42:54.205991 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:42:54 crc kubenswrapper[4805]: E1128 15:42:54.641516 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" podUID="697a2a64-1776-4bb4-95f4-901e1ae2bbc2" Nov 28 15:42:54 crc kubenswrapper[4805]: E1128 15:42:54.641929 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" podUID="abe59027-7c48-4b8e-8255-067032ba364d" Nov 28 15:42:54 crc kubenswrapper[4805]: E1128 15:42:54.641975 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" podUID="f0bf9b46-6d36-47b4-bbac-374d79b553fc" Nov 28 15:42:54 crc kubenswrapper[4805]: E1128 15:42:54.641533 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" podUID="2dac7bca-c6db-40a9-ac64-b46c7722798b" Nov 28 15:42:55 crc kubenswrapper[4805]: I1128 15:42:55.185348 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:55 crc kubenswrapper[4805]: I1128 15:42:55.190545 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dfdaba79-b86b-4a03-87e0-ed9f712ea4bb-cert\") pod \"infra-operator-controller-manager-57548d458d-2p9q6\" (UID: \"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:55 crc kubenswrapper[4805]: I1128 15:42:55.291219 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:42:55 crc kubenswrapper[4805]: I1128 15:42:55.488819 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:55 crc kubenswrapper[4805]: I1128 15:42:55.511489 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cb86467-0536-46c2-a399-53bf3ea553b3-cert\") pod \"openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw\" (UID: \"5cb86467-0536-46c2-a399-53bf3ea553b3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:55 crc kubenswrapper[4805]: I1128 15:42:55.736341 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6"] Nov 28 15:42:55 crc kubenswrapper[4805]: I1128 15:42:55.762546 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:42:55 crc kubenswrapper[4805]: I1128 15:42:55.793393 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:55 crc kubenswrapper[4805]: I1128 15:42:55.793561 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:42:55 crc kubenswrapper[4805]: E1128 15:42:55.793873 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 15:42:55 crc kubenswrapper[4805]: E1128 15:42:55.793953 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:43:11.793935566 +0000 UTC m=+1018.843726877 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "metrics-server-cert" not found Nov 28 15:42:55 crc kubenswrapper[4805]: E1128 15:42:55.793737 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 15:42:55 crc kubenswrapper[4805]: E1128 15:42:55.794612 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs podName:e8c06141-6007-491e-bdd0-6a7654607554 nodeName:}" failed. No retries permitted until 2025-11-28 15:43:11.794583544 +0000 UTC m=+1018.844374855 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs") pod "openstack-operator-controller-manager-6f5f885fb5-4vb9n" (UID: "e8c06141-6007-491e-bdd0-6a7654607554") : secret "webhook-server-cert" not found Nov 28 15:42:55 crc kubenswrapper[4805]: W1128 15:42:55.956826 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfdaba79_b86b_4a03_87e0_ed9f712ea4bb.slice/crio-099db9dde56de165ea18ce61fb7a4d66f6d705c7ff9edc043931390f302c10f7 WatchSource:0}: Error finding container 099db9dde56de165ea18ce61fb7a4d66f6d705c7ff9edc043931390f302c10f7: Status 404 returned error can't find the container with id 099db9dde56de165ea18ce61fb7a4d66f6d705c7ff9edc043931390f302c10f7 Nov 28 15:42:56 crc kubenswrapper[4805]: I1128 15:42:56.616373 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw"] Nov 28 15:42:56 crc kubenswrapper[4805]: I1128 15:42:56.652972 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" event={"ID":"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb","Type":"ContainerStarted","Data":"099db9dde56de165ea18ce61fb7a4d66f6d705c7ff9edc043931390f302c10f7"} Nov 28 15:42:57 crc kubenswrapper[4805]: W1128 15:42:57.360198 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cb86467_0536_46c2_a399_53bf3ea553b3.slice/crio-0cb177c3e3dd1706474f5be78cc6688bde27ca4b4bc3901cfb4dfcee67cd826c WatchSource:0}: Error finding container 0cb177c3e3dd1706474f5be78cc6688bde27ca4b4bc3901cfb4dfcee67cd826c: Status 404 returned error can't find the container with id 0cb177c3e3dd1706474f5be78cc6688bde27ca4b4bc3901cfb4dfcee67cd826c Nov 28 15:42:57 crc kubenswrapper[4805]: I1128 15:42:57.659824 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" event={"ID":"5cb86467-0536-46c2-a399-53bf3ea553b3","Type":"ContainerStarted","Data":"0cb177c3e3dd1706474f5be78cc6688bde27ca4b4bc3901cfb4dfcee67cd826c"} Nov 28 15:42:59 crc kubenswrapper[4805]: I1128 15:42:59.534748 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" Nov 28 15:42:59 crc kubenswrapper[4805]: I1128 15:42:59.822915 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" Nov 28 15:43:00 crc kubenswrapper[4805]: I1128 15:43:00.022852 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" Nov 28 15:43:00 crc kubenswrapper[4805]: I1128 15:43:00.141199 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" Nov 28 15:43:09 crc kubenswrapper[4805]: I1128 15:43:09.758109 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" event={"ID":"abe59027-7c48-4b8e-8255-067032ba364d","Type":"ContainerStarted","Data":"0edf071f6e04e85e248bafe9f15b020269346cb2c9679d2e88ad8326ec8346c9"} Nov 28 15:43:09 crc kubenswrapper[4805]: I1128 15:43:09.759906 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" event={"ID":"cd24b067-2061-47eb-8494-5a531f13b111","Type":"ContainerStarted","Data":"81b54113985c91e65bb19b301fd41d2de3ebe535ae5235c00426a4021b62c055"} Nov 28 15:43:09 crc kubenswrapper[4805]: I1128 15:43:09.761144 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" event={"ID":"dc3a6e21-e0ca-40d0-b140-21215085f97a","Type":"ContainerStarted","Data":"f4b90811821a3e38a9b387fce8c8afd6b7f9c52b23a751f123a45100e036d98b"} Nov 28 15:43:09 crc kubenswrapper[4805]: I1128 15:43:09.762233 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" event={"ID":"3bad1d97-f824-4616-a08d-9e00fd2c5201","Type":"ContainerStarted","Data":"762ecbc0a6eb0525db9228d6937ddd812bb9c654a8df3e4d9be565be6f4d0266"} Nov 28 15:43:09 crc kubenswrapper[4805]: I1128 15:43:09.763705 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" event={"ID":"2dac7bca-c6db-40a9-ac64-b46c7722798b","Type":"ContainerStarted","Data":"0c65758276f0bd46de6c649bd7fcf86d1a0506e0958c3aa5bbb3df9d14de5059"} Nov 28 15:43:09 crc kubenswrapper[4805]: I1128 15:43:09.765173 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" event={"ID":"01f64db7-5c9e-41ea-a595-75dd81511543","Type":"ContainerStarted","Data":"38390c03411090305e1544525a5429affd625a1ee6f2dfbeea062e693ae66773"} Nov 28 15:43:09 crc kubenswrapper[4805]: I1128 15:43:09.766158 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" event={"ID":"519f14b3-c5a1-4bed-8edc-92a84bf0a70d","Type":"ContainerStarted","Data":"4d0229e866e8e46e04cb7ba9bcb66aaa70ce80176c25881f9de0b77c576674b8"} Nov 28 15:43:09 crc kubenswrapper[4805]: I1128 15:43:09.767317 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" event={"ID":"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb","Type":"ContainerStarted","Data":"fcc1f43b8bb8faeb7c7da4fd4e854cb62160f7028a4960f0bf479390982e814e"} Nov 28 15:43:09 crc kubenswrapper[4805]: I1128 15:43:09.768425 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" event={"ID":"8e874a32-dff0-48ad-b43b-a24c364e5e8d","Type":"ContainerStarted","Data":"56b8aa2bca96621a1e8b47ca344da3e050843dd934bcdfeb6e633c2adeb42ac3"} Nov 28 15:43:10 crc kubenswrapper[4805]: I1128 15:43:10.775587 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" Nov 28 15:43:10 crc kubenswrapper[4805]: I1128 15:43:10.777485 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" Nov 28 15:43:10 crc kubenswrapper[4805]: I1128 15:43:10.807161 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2jrp8" podStartSLOduration=10.058399654 podStartE2EDuration="31.807134158s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.198493279 +0000 UTC m=+988.248284590" lastFinishedPulling="2025-11-28 15:43:02.947227773 +0000 UTC m=+1009.997019094" observedRunningTime="2025-11-28 15:43:10.795614791 +0000 UTC m=+1017.845406162" watchObservedRunningTime="2025-11-28 15:43:10.807134158 +0000 UTC m=+1017.856925509" Nov 28 15:43:11 crc kubenswrapper[4805]: I1128 15:43:11.870476 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:43:11 crc kubenswrapper[4805]: I1128 15:43:11.870595 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:43:11 crc kubenswrapper[4805]: I1128 15:43:11.876728 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-metrics-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:43:11 crc kubenswrapper[4805]: I1128 15:43:11.878300 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8c06141-6007-491e-bdd0-6a7654607554-webhook-certs\") pod \"openstack-operator-controller-manager-6f5f885fb5-4vb9n\" (UID: \"e8c06141-6007-491e-bdd0-6a7654607554\") " pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:43:11 crc kubenswrapper[4805]: I1128 15:43:11.949658 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:43:12 crc kubenswrapper[4805]: I1128 15:43:12.416848 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n"] Nov 28 15:43:12 crc kubenswrapper[4805]: I1128 15:43:12.792291 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" event={"ID":"e8c06141-6007-491e-bdd0-6a7654607554","Type":"ContainerStarted","Data":"72e916c01a5575f55a00660a200fb806978a4fd75457043d912908167eb80cdc"} Nov 28 15:43:15 crc kubenswrapper[4805]: I1128 15:43:15.862828 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hnhwx" podStartSLOduration=25.645018099 podStartE2EDuration="36.862791427s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.191996771 +0000 UTC m=+988.241788082" lastFinishedPulling="2025-11-28 15:42:52.409770079 +0000 UTC m=+999.459561410" observedRunningTime="2025-11-28 15:43:15.8552525 +0000 UTC m=+1022.905043851" watchObservedRunningTime="2025-11-28 15:43:15.862791427 +0000 UTC m=+1022.912582808" Nov 28 15:43:16 crc kubenswrapper[4805]: I1128 15:43:16.839610 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" event={"ID":"8ff9654c-7de4-4cc2-af76-863ac6cb8694","Type":"ContainerStarted","Data":"cf30e11ffcd871130378766ef2e34592277d6a68b34d2477f1a4402c2543b0cf"} Nov 28 15:43:16 crc kubenswrapper[4805]: I1128 15:43:16.858198 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-c6lqn" podStartSLOduration=26.818055972 podStartE2EDuration="37.858177579s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.387882835 +0000 UTC m=+988.437674146" lastFinishedPulling="2025-11-28 15:42:52.428004442 +0000 UTC m=+999.477795753" observedRunningTime="2025-11-28 15:43:16.857212993 +0000 UTC m=+1023.907004324" watchObservedRunningTime="2025-11-28 15:43:16.858177579 +0000 UTC m=+1023.907968910" Nov 28 15:43:18 crc kubenswrapper[4805]: I1128 15:43:18.648770 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" event={"ID":"0dc0e269-4d2b-4e4b-afcc-0547ac908a26","Type":"ContainerStarted","Data":"1aa25da9923914033b23d2e7618f071435232f9d580cae00c94321e270324e3a"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.660044 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" event={"ID":"f0bf9b46-6d36-47b4-bbac-374d79b553fc","Type":"ContainerStarted","Data":"5ab611ed9894a5a2baf6de0c2cb0ba09910e8b68dfb523bef8875522cbc3d467"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.661719 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" event={"ID":"446d7f87-a4e9-4a87-9091-926345489a0d","Type":"ContainerStarted","Data":"13a27442f3258ba2c19e5d774c906aabee0aed9799f664dbe7ec74ed05ee8ab4"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.673266 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" event={"ID":"519f14b3-c5a1-4bed-8edc-92a84bf0a70d","Type":"ContainerStarted","Data":"50685d97bbd8c85eae7a84b83225cd2cee43c470727dc422df8534a40dc5e145"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.677789 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" event={"ID":"4a95d51b-3298-4575-84d3-5b1f1174edbe","Type":"ContainerStarted","Data":"f0f316321925086680f124cdbda63c311a3b4944b37ade96d204aa36f0f0c2e9"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.680331 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" event={"ID":"d786fb2a-908c-42bc-8889-f85d48a824c3","Type":"ContainerStarted","Data":"4d4e8a7ddf4650ba940d7944c0e76c02759a5bd2c3c931bb3c61f0a9e61add00"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.682118 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" event={"ID":"6d7afdad-6a62-4c69-b453-dcf5b917a568","Type":"ContainerStarted","Data":"f7678e15d4560547f3d9ea3ff707c013517d1ecf424e6e8c304fefc46585105e"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.683345 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" event={"ID":"25d77f9e-dbcf-42b8-8e8e-28d3a500791d","Type":"ContainerStarted","Data":"72626af7b4222ddda9aa49a63c70c52e96267c98222ae8502c6daacab8ca7b3d"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.684954 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" event={"ID":"5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0","Type":"ContainerStarted","Data":"81dfe4c3d28d9a32176b56b835e5ebe74814b1e1449fe264fb568db6e6e13903"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.686500 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" event={"ID":"75b25042-ef68-47da-873b-6750083ecc7e","Type":"ContainerStarted","Data":"712f8281ff753e7ce95e339ee217d7b839cf7fed08f225850ad4645ffb76e8e5"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.687873 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" event={"ID":"5cb86467-0536-46c2-a399-53bf3ea553b3","Type":"ContainerStarted","Data":"526d1c36bf674e1bd51d5aa9274c1762ca42edeedbdd3aa4b3c7da6ba8f79742"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.689670 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" event={"ID":"697a2a64-1776-4bb4-95f4-901e1ae2bbc2","Type":"ContainerStarted","Data":"522b6a8ed52aa356c49aec34dbd18780045687d0f5c2377c2818f7f2f446d46f"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.691810 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" event={"ID":"49986953-d4ca-4e5f-ac36-f91d86098ff5","Type":"ContainerStarted","Data":"5417eb18056ae2b718cc945d4b98ba4195fe197327d8ab0f80e98153878444db"} Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.692067 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.698217 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" Nov 28 15:43:19 crc kubenswrapper[4805]: I1128 15:43:19.716555 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-p9lc4" podStartSLOduration=12.9239242 podStartE2EDuration="40.716523627s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:40.921683756 +0000 UTC m=+987.971475067" lastFinishedPulling="2025-11-28 15:43:08.714283173 +0000 UTC m=+1015.764074494" observedRunningTime="2025-11-28 15:43:19.714707796 +0000 UTC m=+1026.764499117" watchObservedRunningTime="2025-11-28 15:43:19.716523627 +0000 UTC m=+1026.766314988" Nov 28 15:43:20 crc kubenswrapper[4805]: I1128 15:43:20.703591 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" event={"ID":"8e874a32-dff0-48ad-b43b-a24c364e5e8d","Type":"ContainerStarted","Data":"33a54566daa75d5df62073b4a39d6fd558651fac452c5e68d211af80193b450c"} Nov 28 15:43:20 crc kubenswrapper[4805]: I1128 15:43:20.705926 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" event={"ID":"cd24b067-2061-47eb-8494-5a531f13b111","Type":"ContainerStarted","Data":"e5a2c9d05d6c0cbe6660f0aff44b30947700d161cc1a82991d8c466f5ffe5e61"} Nov 28 15:43:20 crc kubenswrapper[4805]: I1128 15:43:20.710211 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" event={"ID":"dc3a6e21-e0ca-40d0-b140-21215085f97a","Type":"ContainerStarted","Data":"393d9608a707b187cbb10369bb163a25fc0abd4915a5ee1f1e0b2c79b1bb9860"} Nov 28 15:43:20 crc kubenswrapper[4805]: I1128 15:43:20.713178 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" event={"ID":"3bad1d97-f824-4616-a08d-9e00fd2c5201","Type":"ContainerStarted","Data":"7674fb271c638d046e012635bae8f58309810ae4ae7d5edb8ae69eb169c8f4ea"} Nov 28 15:43:20 crc kubenswrapper[4805]: I1128 15:43:20.715817 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" event={"ID":"dfdaba79-b86b-4a03-87e0-ed9f712ea4bb","Type":"ContainerStarted","Data":"2cb15a472b473696f147cbef2a47feb0266f898e505f3b6d798b42dc4f128960"} Nov 28 15:43:20 crc kubenswrapper[4805]: I1128 15:43:20.745308 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" podStartSLOduration=21.493888691 podStartE2EDuration="41.745270848s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:40.973582486 +0000 UTC m=+988.023373797" lastFinishedPulling="2025-11-28 15:43:01.224964633 +0000 UTC m=+1008.274755954" observedRunningTime="2025-11-28 15:43:20.743073227 +0000 UTC m=+1027.792864548" watchObservedRunningTime="2025-11-28 15:43:20.745270848 +0000 UTC m=+1027.795062169" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.730821 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" event={"ID":"e8c06141-6007-491e-bdd0-6a7654607554","Type":"ContainerStarted","Data":"11f1816c614b83a0d2ee8dc877907139d17bf5e6a10e267071ce86aad7c921ef"} Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.731626 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.731898 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.732386 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.735751 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.736065 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.736968 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.757891 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-skdkr" podStartSLOduration=15.890627329 podStartE2EDuration="42.757856643s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:40.979539229 +0000 UTC m=+988.029330550" lastFinishedPulling="2025-11-28 15:43:07.846768513 +0000 UTC m=+1014.896559864" observedRunningTime="2025-11-28 15:43:21.748846165 +0000 UTC m=+1028.798637476" watchObservedRunningTime="2025-11-28 15:43:21.757856643 +0000 UTC m=+1028.807648004" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.781950 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-m6cln" podStartSLOduration=15.180211595 podStartE2EDuration="42.781923206s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:40.634828416 +0000 UTC m=+987.684619727" lastFinishedPulling="2025-11-28 15:43:08.236540007 +0000 UTC m=+1015.286331338" observedRunningTime="2025-11-28 15:43:21.769481614 +0000 UTC m=+1028.819272935" watchObservedRunningTime="2025-11-28 15:43:21.781923206 +0000 UTC m=+1028.831714537" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.796065 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-v7f97" podStartSLOduration=31.314482961 podStartE2EDuration="42.796048485s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:40.961276906 +0000 UTC m=+988.011068217" lastFinishedPulling="2025-11-28 15:42:52.44284243 +0000 UTC m=+999.492633741" observedRunningTime="2025-11-28 15:43:21.78640944 +0000 UTC m=+1028.836200781" watchObservedRunningTime="2025-11-28 15:43:21.796048485 +0000 UTC m=+1028.845839796" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.821383 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d77b94747-mg6cr" podStartSLOduration=15.414896049 podStartE2EDuration="42.821345982s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.198567101 +0000 UTC m=+988.248358432" lastFinishedPulling="2025-11-28 15:43:08.605017054 +0000 UTC m=+1015.654808365" observedRunningTime="2025-11-28 15:43:21.808529828 +0000 UTC m=+1028.858321149" watchObservedRunningTime="2025-11-28 15:43:21.821345982 +0000 UTC m=+1028.871137313" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.860133 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-m9thn" podStartSLOduration=14.566882524 podStartE2EDuration="41.860075468s" podCreationTimestamp="2025-11-28 15:42:40 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.410993001 +0000 UTC m=+988.460784312" lastFinishedPulling="2025-11-28 15:43:08.704185945 +0000 UTC m=+1015.753977256" observedRunningTime="2025-11-28 15:43:21.854300659 +0000 UTC m=+1028.904091980" watchObservedRunningTime="2025-11-28 15:43:21.860075468 +0000 UTC m=+1028.909866779" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.890825 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" podStartSLOduration=15.236414945 podStartE2EDuration="42.890802395s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.018313137 +0000 UTC m=+988.068104448" lastFinishedPulling="2025-11-28 15:43:08.672700587 +0000 UTC m=+1015.722491898" observedRunningTime="2025-11-28 15:43:21.889306753 +0000 UTC m=+1028.939098074" watchObservedRunningTime="2025-11-28 15:43:21.890802395 +0000 UTC m=+1028.940593706" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.908985 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" podStartSLOduration=16.084948901 podStartE2EDuration="42.908970074s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.41202977 +0000 UTC m=+988.461821081" lastFinishedPulling="2025-11-28 15:43:08.236050923 +0000 UTC m=+1015.285842254" observedRunningTime="2025-11-28 15:43:21.904276136 +0000 UTC m=+1028.954067447" watchObservedRunningTime="2025-11-28 15:43:21.908970074 +0000 UTC m=+1028.958761385" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.922673 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" podStartSLOduration=15.390410344 podStartE2EDuration="42.922656941s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.198595482 +0000 UTC m=+988.248386793" lastFinishedPulling="2025-11-28 15:43:08.730842079 +0000 UTC m=+1015.780633390" observedRunningTime="2025-11-28 15:43:21.921517861 +0000 UTC m=+1028.971309202" watchObservedRunningTime="2025-11-28 15:43:21.922656941 +0000 UTC m=+1028.972448242" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.949289 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" podStartSLOduration=23.172664813 podStartE2EDuration="42.949268715s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.430340384 +0000 UTC m=+988.480131695" lastFinishedPulling="2025-11-28 15:43:01.206944246 +0000 UTC m=+1008.256735597" observedRunningTime="2025-11-28 15:43:21.944799501 +0000 UTC m=+1028.994590812" watchObservedRunningTime="2025-11-28 15:43:21.949268715 +0000 UTC m=+1028.999060026" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.960187 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" podStartSLOduration=14.87605129 podStartE2EDuration="42.960171865s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:40.587126193 +0000 UTC m=+987.636917504" lastFinishedPulling="2025-11-28 15:43:08.671246768 +0000 UTC m=+1015.721038079" observedRunningTime="2025-11-28 15:43:21.960076953 +0000 UTC m=+1029.009868274" watchObservedRunningTime="2025-11-28 15:43:21.960171865 +0000 UTC m=+1029.009963176" Nov 28 15:43:21 crc kubenswrapper[4805]: I1128 15:43:21.978654 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jxp" podStartSLOduration=31.152208602 podStartE2EDuration="42.978631033s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:40.629922891 +0000 UTC m=+987.679714202" lastFinishedPulling="2025-11-28 15:42:52.456345322 +0000 UTC m=+999.506136633" observedRunningTime="2025-11-28 15:43:21.972309739 +0000 UTC m=+1029.022101070" watchObservedRunningTime="2025-11-28 15:43:21.978631033 +0000 UTC m=+1029.028422344" Nov 28 15:43:22 crc kubenswrapper[4805]: I1128 15:43:22.029852 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" podStartSLOduration=15.770674968 podStartE2EDuration="43.029829823s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:40.976711352 +0000 UTC m=+988.026502673" lastFinishedPulling="2025-11-28 15:43:08.235866217 +0000 UTC m=+1015.285657528" observedRunningTime="2025-11-28 15:43:22.025026401 +0000 UTC m=+1029.074817722" watchObservedRunningTime="2025-11-28 15:43:22.029829823 +0000 UTC m=+1029.079621134" Nov 28 15:43:22 crc kubenswrapper[4805]: I1128 15:43:22.072727 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" podStartSLOduration=15.878710112 podStartE2EDuration="43.072706494s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.410849667 +0000 UTC m=+988.460640978" lastFinishedPulling="2025-11-28 15:43:08.604846049 +0000 UTC m=+1015.654637360" observedRunningTime="2025-11-28 15:43:22.070122463 +0000 UTC m=+1029.119913774" watchObservedRunningTime="2025-11-28 15:43:22.072706494 +0000 UTC m=+1029.122497815" Nov 28 15:43:22 crc kubenswrapper[4805]: I1128 15:43:22.096988 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" podStartSLOduration=23.246224517 podStartE2EDuration="43.096954581s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.391036492 +0000 UTC m=+988.440827803" lastFinishedPulling="2025-11-28 15:43:01.241766536 +0000 UTC m=+1008.291557867" observedRunningTime="2025-11-28 15:43:22.095664726 +0000 UTC m=+1029.145456037" watchObservedRunningTime="2025-11-28 15:43:22.096954581 +0000 UTC m=+1029.146745892" Nov 28 15:43:22 crc kubenswrapper[4805]: I1128 15:43:22.120836 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" podStartSLOduration=15.447698563 podStartE2EDuration="43.120814169s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.021582277 +0000 UTC m=+988.071373588" lastFinishedPulling="2025-11-28 15:43:08.694697893 +0000 UTC m=+1015.744489194" observedRunningTime="2025-11-28 15:43:22.10887687 +0000 UTC m=+1029.158668181" watchObservedRunningTime="2025-11-28 15:43:22.120814169 +0000 UTC m=+1029.170605480" Nov 28 15:43:22 crc kubenswrapper[4805]: I1128 15:43:22.755321 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" podStartSLOduration=38.492062726 podStartE2EDuration="43.755304822s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:55.961746318 +0000 UTC m=+1003.011537619" lastFinishedPulling="2025-11-28 15:43:01.224988394 +0000 UTC m=+1008.274779715" observedRunningTime="2025-11-28 15:43:22.751819316 +0000 UTC m=+1029.801610627" watchObservedRunningTime="2025-11-28 15:43:22.755304822 +0000 UTC m=+1029.805096133" Nov 28 15:43:22 crc kubenswrapper[4805]: I1128 15:43:22.776630 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" podStartSLOduration=43.776611579 podStartE2EDuration="43.776611579s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:43:22.774503731 +0000 UTC m=+1029.824295062" watchObservedRunningTime="2025-11-28 15:43:22.776611579 +0000 UTC m=+1029.826402890" Nov 28 15:43:23 crc kubenswrapper[4805]: I1128 15:43:23.746664 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" event={"ID":"5cb86467-0536-46c2-a399-53bf3ea553b3","Type":"ContainerStarted","Data":"616d5cbc2a678942d2a307cb5fd3ef52e14ff6abff14271b43bee11b3a351ed9"} Nov 28 15:43:24 crc kubenswrapper[4805]: I1128 15:43:24.754490 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:43:24 crc kubenswrapper[4805]: I1128 15:43:24.762508 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" Nov 28 15:43:24 crc kubenswrapper[4805]: I1128 15:43:24.800379 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw" podStartSLOduration=34.468852721 podStartE2EDuration="45.800347801s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:57.365007363 +0000 UTC m=+1004.414798674" lastFinishedPulling="2025-11-28 15:43:08.696502443 +0000 UTC m=+1015.746293754" observedRunningTime="2025-11-28 15:43:24.796565297 +0000 UTC m=+1031.846356608" watchObservedRunningTime="2025-11-28 15:43:24.800347801 +0000 UTC m=+1031.850139102" Nov 28 15:43:25 crc kubenswrapper[4805]: I1128 15:43:25.291453 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:43:25 crc kubenswrapper[4805]: I1128 15:43:25.298320 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2p9q6" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.514020 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.516298 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-lh9jk" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.614392 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.617836 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-jl2bq" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.675843 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.681314 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jn44r" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.912255 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.914202 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-5m48r" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.931836 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.933564 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bqmpw" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.968079 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" Nov 28 15:43:29 crc kubenswrapper[4805]: I1128 15:43:29.970973 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-lkqjk" Nov 28 15:43:30 crc kubenswrapper[4805]: I1128 15:43:30.008719 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" Nov 28 15:43:30 crc kubenswrapper[4805]: I1128 15:43:30.016525 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-2wbp7" Nov 28 15:43:30 crc kubenswrapper[4805]: I1128 15:43:30.133073 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" Nov 28 15:43:30 crc kubenswrapper[4805]: I1128 15:43:30.136031 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-pmlh7" Nov 28 15:43:30 crc kubenswrapper[4805]: I1128 15:43:30.185011 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" Nov 28 15:43:30 crc kubenswrapper[4805]: I1128 15:43:30.186593 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9q29v" Nov 28 15:43:30 crc kubenswrapper[4805]: I1128 15:43:30.480352 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" Nov 28 15:43:30 crc kubenswrapper[4805]: I1128 15:43:30.482725 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2pnsl" Nov 28 15:43:31 crc kubenswrapper[4805]: I1128 15:43:31.951004 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:43:31 crc kubenswrapper[4805]: I1128 15:43:31.957312 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6f5f885fb5-4vb9n" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.582296 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-557f57d995-xd9gt"] Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.584908 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.587994 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-njvr5" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.588475 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.588624 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.592145 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-xd9gt"] Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.593816 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.599025 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86b55ee-3e73-4509-bf86-66d380dbb8d0-config\") pod \"dnsmasq-dns-557f57d995-xd9gt\" (UID: \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\") " pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.599136 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnnhc\" (UniqueName: \"kubernetes.io/projected/b86b55ee-3e73-4509-bf86-66d380dbb8d0-kube-api-access-lnnhc\") pod \"dnsmasq-dns-557f57d995-xd9gt\" (UID: \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\") " pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.663730 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-s67c9"] Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.665807 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.669596 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.671116 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-s67c9"] Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.701201 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnnhc\" (UniqueName: \"kubernetes.io/projected/b86b55ee-3e73-4509-bf86-66d380dbb8d0-kube-api-access-lnnhc\") pod \"dnsmasq-dns-557f57d995-xd9gt\" (UID: \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\") " pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.701273 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86b55ee-3e73-4509-bf86-66d380dbb8d0-config\") pod \"dnsmasq-dns-557f57d995-xd9gt\" (UID: \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\") " pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.702138 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86b55ee-3e73-4509-bf86-66d380dbb8d0-config\") pod \"dnsmasq-dns-557f57d995-xd9gt\" (UID: \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\") " pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.729532 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnnhc\" (UniqueName: \"kubernetes.io/projected/b86b55ee-3e73-4509-bf86-66d380dbb8d0-kube-api-access-lnnhc\") pod \"dnsmasq-dns-557f57d995-xd9gt\" (UID: \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\") " pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.803141 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-config\") pod \"dnsmasq-dns-766fdc659c-s67c9\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.803257 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8j6l\" (UniqueName: \"kubernetes.io/projected/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-kube-api-access-n8j6l\") pod \"dnsmasq-dns-766fdc659c-s67c9\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.803322 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-dns-svc\") pod \"dnsmasq-dns-766fdc659c-s67c9\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.905066 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8j6l\" (UniqueName: \"kubernetes.io/projected/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-kube-api-access-n8j6l\") pod \"dnsmasq-dns-766fdc659c-s67c9\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.905166 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-dns-svc\") pod \"dnsmasq-dns-766fdc659c-s67c9\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.905196 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-config\") pod \"dnsmasq-dns-766fdc659c-s67c9\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.906163 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-config\") pod \"dnsmasq-dns-766fdc659c-s67c9\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.906200 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-dns-svc\") pod \"dnsmasq-dns-766fdc659c-s67c9\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.912658 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.936543 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8j6l\" (UniqueName: \"kubernetes.io/projected/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-kube-api-access-n8j6l\") pod \"dnsmasq-dns-766fdc659c-s67c9\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:49 crc kubenswrapper[4805]: I1128 15:43:49.983460 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:43:50 crc kubenswrapper[4805]: I1128 15:43:50.395310 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-xd9gt"] Nov 28 15:43:50 crc kubenswrapper[4805]: I1128 15:43:50.446474 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-s67c9"] Nov 28 15:43:50 crc kubenswrapper[4805]: W1128 15:43:50.455215 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0f37808_4526_4e50_ba7f_8a0974ae0bf3.slice/crio-b632c02467f203642764b049ee266848a86a725d50462b7d128911eebfc4f333 WatchSource:0}: Error finding container b632c02467f203642764b049ee266848a86a725d50462b7d128911eebfc4f333: Status 404 returned error can't find the container with id b632c02467f203642764b049ee266848a86a725d50462b7d128911eebfc4f333 Nov 28 15:43:50 crc kubenswrapper[4805]: I1128 15:43:50.971334 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766fdc659c-s67c9" event={"ID":"d0f37808-4526-4e50-ba7f-8a0974ae0bf3","Type":"ContainerStarted","Data":"b632c02467f203642764b049ee266848a86a725d50462b7d128911eebfc4f333"} Nov 28 15:43:50 crc kubenswrapper[4805]: I1128 15:43:50.972635 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-557f57d995-xd9gt" event={"ID":"b86b55ee-3e73-4509-bf86-66d380dbb8d0","Type":"ContainerStarted","Data":"9613c43646bfca34fa76accd09288ef0c0466c93cbde786e8ef93fd02196e39f"} Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.026981 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-xd9gt"] Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.051424 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-q6hk5"] Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.052676 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.062387 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-q6hk5"] Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.241945 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-config\") pod \"dnsmasq-dns-57dc4c6697-q6hk5\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.242057 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-dns-svc\") pod \"dnsmasq-dns-57dc4c6697-q6hk5\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.242118 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrzn5\" (UniqueName: \"kubernetes.io/projected/7c7eed7b-132f-4796-bcc7-fb991373bded-kube-api-access-jrzn5\") pod \"dnsmasq-dns-57dc4c6697-q6hk5\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.296012 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-s67c9"] Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.321036 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b5zbf"] Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.322188 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.331175 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b5zbf"] Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.343081 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-dns-svc\") pod \"dnsmasq-dns-57dc4c6697-q6hk5\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.343171 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrzn5\" (UniqueName: \"kubernetes.io/projected/7c7eed7b-132f-4796-bcc7-fb991373bded-kube-api-access-jrzn5\") pod \"dnsmasq-dns-57dc4c6697-q6hk5\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.343209 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-config\") pod \"dnsmasq-dns-57dc4c6697-q6hk5\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.343978 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-config\") pod \"dnsmasq-dns-57dc4c6697-q6hk5\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.344582 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-dns-svc\") pod \"dnsmasq-dns-57dc4c6697-q6hk5\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.372816 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrzn5\" (UniqueName: \"kubernetes.io/projected/7c7eed7b-132f-4796-bcc7-fb991373bded-kube-api-access-jrzn5\") pod \"dnsmasq-dns-57dc4c6697-q6hk5\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.385275 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.444978 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-dns-svc\") pod \"dnsmasq-dns-8446fd7c75-b5zbf\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.445298 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-config\") pod \"dnsmasq-dns-8446fd7c75-b5zbf\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.445604 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrv4c\" (UniqueName: \"kubernetes.io/projected/86255190-b7c6-4ede-8327-196ffbceb9d0-kube-api-access-rrv4c\") pod \"dnsmasq-dns-8446fd7c75-b5zbf\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.547802 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-dns-svc\") pod \"dnsmasq-dns-8446fd7c75-b5zbf\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.547886 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-config\") pod \"dnsmasq-dns-8446fd7c75-b5zbf\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.547935 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrv4c\" (UniqueName: \"kubernetes.io/projected/86255190-b7c6-4ede-8327-196ffbceb9d0-kube-api-access-rrv4c\") pod \"dnsmasq-dns-8446fd7c75-b5zbf\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.548787 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-dns-svc\") pod \"dnsmasq-dns-8446fd7c75-b5zbf\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.549826 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-config\") pod \"dnsmasq-dns-8446fd7c75-b5zbf\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.563393 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrv4c\" (UniqueName: \"kubernetes.io/projected/86255190-b7c6-4ede-8327-196ffbceb9d0-kube-api-access-rrv4c\") pod \"dnsmasq-dns-8446fd7c75-b5zbf\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.638745 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.912397 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-q6hk5"] Nov 28 15:43:52 crc kubenswrapper[4805]: W1128 15:43:52.925406 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c7eed7b_132f_4796_bcc7_fb991373bded.slice/crio-a2ece9cca3f41933f7e7a4fbd0348c83d059b5706d9f3ffe3ce31f24fd48f987 WatchSource:0}: Error finding container a2ece9cca3f41933f7e7a4fbd0348c83d059b5706d9f3ffe3ce31f24fd48f987: Status 404 returned error can't find the container with id a2ece9cca3f41933f7e7a4fbd0348c83d059b5706d9f3ffe3ce31f24fd48f987 Nov 28 15:43:52 crc kubenswrapper[4805]: I1128 15:43:52.999818 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" event={"ID":"7c7eed7b-132f-4796-bcc7-fb991373bded","Type":"ContainerStarted","Data":"a2ece9cca3f41933f7e7a4fbd0348c83d059b5706d9f3ffe3ce31f24fd48f987"} Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.168715 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b5zbf"] Nov 28 15:43:53 crc kubenswrapper[4805]: W1128 15:43:53.177435 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86255190_b7c6_4ede_8327_196ffbceb9d0.slice/crio-8e4421410f29d8388091403fae1550f746fbfad3dadc07432d7d71155543a25c WatchSource:0}: Error finding container 8e4421410f29d8388091403fae1550f746fbfad3dadc07432d7d71155543a25c: Status 404 returned error can't find the container with id 8e4421410f29d8388091403fae1550f746fbfad3dadc07432d7d71155543a25c Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.196335 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.198989 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.201603 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.201710 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.201603 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.201958 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.202056 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.202066 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-sbc88" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.202999 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.264307 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.363860 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjslc\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-kube-api-access-wjslc\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364196 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364220 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364242 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364278 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364344 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364387 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364405 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364424 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364466 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.364503 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.445982 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.447174 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.449821 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.449888 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.449930 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.449825 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.449883 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.450412 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.451333 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-8869p" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.460570 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465559 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465598 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465626 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465676 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465704 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465724 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465744 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465765 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465784 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465801 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjslc\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-kube-api-access-wjslc\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.465826 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.466435 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.466698 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.467400 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.468197 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.468703 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.468723 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.478040 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.478249 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.480175 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.496009 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjslc\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-kube-api-access-wjslc\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.496445 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.502109 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.537764 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.566935 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzjl9\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-kube-api-access-kzjl9\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.566983 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.567017 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.567035 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.567058 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.567075 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.567092 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.567108 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.567126 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.567173 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb122aae-0e09-46b2-926c-037d25e79477-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.567188 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb122aae-0e09-46b2-926c-037d25e79477-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668521 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb122aae-0e09-46b2-926c-037d25e79477-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668566 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb122aae-0e09-46b2-926c-037d25e79477-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668618 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzjl9\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-kube-api-access-kzjl9\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668650 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668683 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668706 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668736 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668761 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668786 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668816 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.668848 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.669554 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.670381 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.671331 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.671668 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.672757 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.673467 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.676482 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb122aae-0e09-46b2-926c-037d25e79477-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.676804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.676981 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.679809 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb122aae-0e09-46b2-926c-037d25e79477-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.684805 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzjl9\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-kube-api-access-kzjl9\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.695713 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:53 crc kubenswrapper[4805]: I1128 15:43:53.771823 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.022686 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" event={"ID":"86255190-b7c6-4ede-8327-196ffbceb9d0","Type":"ContainerStarted","Data":"8e4421410f29d8388091403fae1550f746fbfad3dadc07432d7d71155543a25c"} Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.058494 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 15:43:54 crc kubenswrapper[4805]: W1128 15:43:54.067995 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b51e8d9_5cd6_42f7_9fbf_ff2c6d856c7a.slice/crio-d553f54295d1ab3f97b6a308ebac8e0039b9aaecbf4c5e8b9ced6cffda3fc156 WatchSource:0}: Error finding container d553f54295d1ab3f97b6a308ebac8e0039b9aaecbf4c5e8b9ced6cffda3fc156: Status 404 returned error can't find the container with id d553f54295d1ab3f97b6a308ebac8e0039b9aaecbf4c5e8b9ced6cffda3fc156 Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.301786 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.978143 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.981282 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.982966 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.984263 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.984723 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.988383 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-8kzms" Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.989846 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 28 15:43:54 crc kubenswrapper[4805]: I1128 15:43:54.998249 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.035589 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb122aae-0e09-46b2-926c-037d25e79477","Type":"ContainerStarted","Data":"90c5973fc0ce456c98a3efe57e0b031cf0efbc7f1b4525d3b71703c53e2ace07"} Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.037155 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a","Type":"ContainerStarted","Data":"d553f54295d1ab3f97b6a308ebac8e0039b9aaecbf4c5e8b9ced6cffda3fc156"} Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.106840 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.106889 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-default\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.106922 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kolla-config\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.106967 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.106986 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.107008 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.107045 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rwzl\" (UniqueName: \"kubernetes.io/projected/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kube-api-access-6rwzl\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.107097 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.208730 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.208780 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.208855 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-default\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.209069 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kolla-config\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.209219 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.209240 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.209289 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.209345 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rwzl\" (UniqueName: \"kubernetes.io/projected/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kube-api-access-6rwzl\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.214946 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kolla-config\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.215194 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.219798 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.220751 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-default\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.221298 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.237857 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.242905 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.248065 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rwzl\" (UniqueName: \"kubernetes.io/projected/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kube-api-access-6rwzl\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.341241 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " pod="openstack/openstack-galera-0" Nov 28 15:43:55 crc kubenswrapper[4805]: I1128 15:43:55.607125 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.317600 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.323387 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.327415 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.327811 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.328111 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.328323 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-zcd9z" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.344891 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.437657 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.437805 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.437833 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.437893 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smwfr\" (UniqueName: \"kubernetes.io/projected/452c347f-4ee3-46de-ba8f-c83300966f5d-kube-api-access-smwfr\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.437951 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.438729 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.438780 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.438967 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.540710 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.540787 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.540825 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.540856 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.540878 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smwfr\" (UniqueName: \"kubernetes.io/projected/452c347f-4ee3-46de-ba8f-c83300966f5d-kube-api-access-smwfr\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.540902 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.540961 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.540996 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.541494 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.541586 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.542788 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.543677 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.544072 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.548537 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.560278 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.564370 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smwfr\" (UniqueName: \"kubernetes.io/projected/452c347f-4ee3-46de-ba8f-c83300966f5d-kube-api-access-smwfr\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.565117 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.653515 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.668530 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.675242 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.677111 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hnhv4" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.677174 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.677546 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.681463 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.845096 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-config-data\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.845197 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-combined-ca-bundle\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.845225 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmdz7\" (UniqueName: \"kubernetes.io/projected/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kube-api-access-nmdz7\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.845518 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-memcached-tls-certs\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.845679 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kolla-config\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.947454 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-config-data\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.947539 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-combined-ca-bundle\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.947562 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmdz7\" (UniqueName: \"kubernetes.io/projected/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kube-api-access-nmdz7\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.947602 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-memcached-tls-certs\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.947650 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kolla-config\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.948641 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-config-data\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.948837 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kolla-config\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.954164 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-memcached-tls-certs\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.963811 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-combined-ca-bundle\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:56 crc kubenswrapper[4805]: I1128 15:43:56.984108 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmdz7\" (UniqueName: \"kubernetes.io/projected/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kube-api-access-nmdz7\") pod \"memcached-0\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " pod="openstack/memcached-0" Nov 28 15:43:57 crc kubenswrapper[4805]: I1128 15:43:57.006696 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 15:43:58 crc kubenswrapper[4805]: I1128 15:43:58.641053 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:43:58 crc kubenswrapper[4805]: I1128 15:43:58.642242 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 15:43:58 crc kubenswrapper[4805]: I1128 15:43:58.644644 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-58ndb" Nov 28 15:43:58 crc kubenswrapper[4805]: I1128 15:43:58.661215 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:43:58 crc kubenswrapper[4805]: I1128 15:43:58.778411 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkhd6\" (UniqueName: \"kubernetes.io/projected/fb35eb4d-a935-4c99-a3f9-e2fc22c32157-kube-api-access-tkhd6\") pod \"kube-state-metrics-0\" (UID: \"fb35eb4d-a935-4c99-a3f9-e2fc22c32157\") " pod="openstack/kube-state-metrics-0" Nov 28 15:43:58 crc kubenswrapper[4805]: I1128 15:43:58.879936 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkhd6\" (UniqueName: \"kubernetes.io/projected/fb35eb4d-a935-4c99-a3f9-e2fc22c32157-kube-api-access-tkhd6\") pod \"kube-state-metrics-0\" (UID: \"fb35eb4d-a935-4c99-a3f9-e2fc22c32157\") " pod="openstack/kube-state-metrics-0" Nov 28 15:43:58 crc kubenswrapper[4805]: I1128 15:43:58.896505 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkhd6\" (UniqueName: \"kubernetes.io/projected/fb35eb4d-a935-4c99-a3f9-e2fc22c32157-kube-api-access-tkhd6\") pod \"kube-state-metrics-0\" (UID: \"fb35eb4d-a935-4c99-a3f9-e2fc22c32157\") " pod="openstack/kube-state-metrics-0" Nov 28 15:43:59 crc kubenswrapper[4805]: I1128 15:43:59.015617 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.180574 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.181820 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.185090 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.186999 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-l8pfb" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.193525 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.194896 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.194904 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.195281 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.365035 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-config\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.365287 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.365426 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwlbk\" (UniqueName: \"kubernetes.io/projected/bee11cf6-ec1b-40bb-8030-044d8f06591e-kube-api-access-fwlbk\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.365567 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.365720 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.365876 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.366027 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.366154 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.468230 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-config\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.468741 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.468940 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwlbk\" (UniqueName: \"kubernetes.io/projected/bee11cf6-ec1b-40bb-8030-044d8f06591e-kube-api-access-fwlbk\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.469125 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.469412 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.469777 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.470089 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.470467 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.470755 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.470844 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.473559 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.474537 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-config\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.478634 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.478911 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.479401 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.505228 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.508321 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwlbk\" (UniqueName: \"kubernetes.io/projected/bee11cf6-ec1b-40bb-8030-044d8f06591e-kube-api-access-fwlbk\") pod \"ovsdbserver-nb-0\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.527004 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.742873 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4q9c2"] Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.744441 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.753205 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.753461 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-g6njd" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.757984 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-vssvf"] Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.759348 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.759556 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.783630 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4q9c2"] Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.791682 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vssvf"] Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.890803 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d793e5-c2d1-4630-95ff-615cd30b5d04-scripts\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891141 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2v8k\" (UniqueName: \"kubernetes.io/projected/89d793e5-c2d1-4630-95ff-615cd30b5d04-kube-api-access-g2v8k\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891173 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-combined-ca-bundle\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891193 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-run\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891209 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-log-ovn\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891226 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjhkb\" (UniqueName: \"kubernetes.io/projected/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-kube-api-access-jjhkb\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891248 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-ovn-controller-tls-certs\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891264 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-lib\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891285 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-etc-ovs\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891348 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-scripts\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891407 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run-ovn\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891427 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-log\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.891461 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.992769 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-scripts\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.992829 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run-ovn\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.992863 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-log\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.992904 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.992957 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d793e5-c2d1-4630-95ff-615cd30b5d04-scripts\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.992984 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2v8k\" (UniqueName: \"kubernetes.io/projected/89d793e5-c2d1-4630-95ff-615cd30b5d04-kube-api-access-g2v8k\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993015 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-combined-ca-bundle\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993037 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-run\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993055 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-log-ovn\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993071 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjhkb\" (UniqueName: \"kubernetes.io/projected/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-kube-api-access-jjhkb\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993090 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-ovn-controller-tls-certs\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993107 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-lib\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993129 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-etc-ovs\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993467 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run-ovn\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993504 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-etc-ovs\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993563 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-log\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993591 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-run\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993902 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-log-ovn\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.993959 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.994085 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-lib\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.995551 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d793e5-c2d1-4630-95ff-615cd30b5d04-scripts\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.997281 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-scripts\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.998054 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-ovn-controller-tls-certs\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:03 crc kubenswrapper[4805]: I1128 15:44:03.999200 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-combined-ca-bundle\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:04 crc kubenswrapper[4805]: I1128 15:44:04.011533 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjhkb\" (UniqueName: \"kubernetes.io/projected/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-kube-api-access-jjhkb\") pod \"ovn-controller-4q9c2\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:04 crc kubenswrapper[4805]: I1128 15:44:04.016764 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2v8k\" (UniqueName: \"kubernetes.io/projected/89d793e5-c2d1-4630-95ff-615cd30b5d04-kube-api-access-g2v8k\") pod \"ovn-controller-ovs-vssvf\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:04 crc kubenswrapper[4805]: I1128 15:44:04.113942 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:04 crc kubenswrapper[4805]: I1128 15:44:04.255268 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.477180 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.479967 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.482283 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.482878 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.483335 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.485288 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jcd56" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.487227 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.632809 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.632864 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.632890 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsnx2\" (UniqueName: \"kubernetes.io/projected/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-kube-api-access-xsnx2\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.632931 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.633053 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.633120 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.633453 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.633567 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-config\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.735712 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.735769 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-config\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.735835 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.735861 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.735884 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsnx2\" (UniqueName: \"kubernetes.io/projected/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-kube-api-access-xsnx2\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.735920 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.735974 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.736018 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.737404 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.737434 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.737784 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.738051 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-config\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.742261 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.743281 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.746676 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.766898 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsnx2\" (UniqueName: \"kubernetes.io/projected/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-kube-api-access-xsnx2\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.773529 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:06 crc kubenswrapper[4805]: I1128 15:44:06.797503 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:11 crc kubenswrapper[4805]: I1128 15:44:11.060639 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:44:11 crc kubenswrapper[4805]: I1128 15:44:11.061427 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.395296 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.395850 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jrzn5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57dc4c6697-q6hk5_openstack(7c7eed7b-132f-4796-bcc7-fb991373bded): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.397373 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" podUID="7c7eed7b-132f-4796-bcc7-fb991373bded" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.409088 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.409250 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rrv4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-8446fd7c75-b5zbf_openstack(86255190-b7c6-4ede-8327-196ffbceb9d0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.410634 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" podUID="86255190-b7c6-4ede-8327-196ffbceb9d0" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.419664 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.419841 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n8j6l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-766fdc659c-s67c9_openstack(d0f37808-4526-4e50-ba7f-8a0974ae0bf3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.421032 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-766fdc659c-s67c9" podUID="d0f37808-4526-4e50-ba7f-8a0974ae0bf3" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.451280 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.451429 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lnnhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-557f57d995-xd9gt_openstack(b86b55ee-3e73-4509-bf86-66d380dbb8d0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 15:44:15 crc kubenswrapper[4805]: E1128 15:44:15.452608 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-557f57d995-xd9gt" podUID="b86b55ee-3e73-4509-bf86-66d380dbb8d0" Nov 28 15:44:15 crc kubenswrapper[4805]: I1128 15:44:15.874101 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:44:15 crc kubenswrapper[4805]: W1128 15:44:15.876581 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb35eb4d_a935_4c99_a3f9_e2fc22c32157.slice/crio-d9bc28c0f716c3c13619ee4f2c3427c35314207f620bdbb23394af50f502b1c3 WatchSource:0}: Error finding container d9bc28c0f716c3c13619ee4f2c3427c35314207f620bdbb23394af50f502b1c3: Status 404 returned error can't find the container with id d9bc28c0f716c3c13619ee4f2c3427c35314207f620bdbb23394af50f502b1c3 Nov 28 15:44:15 crc kubenswrapper[4805]: I1128 15:44:15.904048 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 28 15:44:15 crc kubenswrapper[4805]: W1128 15:44:15.905372 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40e151fc_3a1e_4b10_8a6a_59bbb17b8d29.slice/crio-925b22ec00321cc888cc9fbf692470a9a21f58a9e31bdea13424e9a4e8bc0f0d WatchSource:0}: Error finding container 925b22ec00321cc888cc9fbf692470a9a21f58a9e31bdea13424e9a4e8bc0f0d: Status 404 returned error can't find the container with id 925b22ec00321cc888cc9fbf692470a9a21f58a9e31bdea13424e9a4e8bc0f0d Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.045931 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 15:44:16 crc kubenswrapper[4805]: W1128 15:44:16.049473 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5243b85f_7ec0_4cca_9f10_de8b40b0a0f4.slice/crio-ef34273fa04965fe17f2bbd3aa0bf4f1851554d8ea6815f2445e709987a916ff WatchSource:0}: Error finding container ef34273fa04965fe17f2bbd3aa0bf4f1851554d8ea6815f2445e709987a916ff: Status 404 returned error can't find the container with id ef34273fa04965fe17f2bbd3aa0bf4f1851554d8ea6815f2445e709987a916ff Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.054698 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.064223 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4q9c2"] Nov 28 15:44:16 crc kubenswrapper[4805]: W1128 15:44:16.085981 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbee11cf6_ec1b_40bb_8030_044d8f06591e.slice/crio-fa7415ffdccf3dfa7986ea148ea45e4b64a565d2af7811fce09af66a733bab92 WatchSource:0}: Error finding container fa7415ffdccf3dfa7986ea148ea45e4b64a565d2af7811fce09af66a733bab92: Status 404 returned error can't find the container with id fa7415ffdccf3dfa7986ea148ea45e4b64a565d2af7811fce09af66a733bab92 Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.092778 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.154488 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 15:44:16 crc kubenswrapper[4805]: W1128 15:44:16.159050 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07201b49_2b7c_4e44_a6a1_3a2dec9df8bd.slice/crio-3a331050dd668115aeeac6c11892c1f6cc093d8d9d3b5acc42ae42790c6e35e0 WatchSource:0}: Error finding container 3a331050dd668115aeeac6c11892c1f6cc093d8d9d3b5acc42ae42790c6e35e0: Status 404 returned error can't find the container with id 3a331050dd668115aeeac6c11892c1f6cc093d8d9d3b5acc42ae42790c6e35e0 Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.196823 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fb35eb4d-a935-4c99-a3f9-e2fc22c32157","Type":"ContainerStarted","Data":"d9bc28c0f716c3c13619ee4f2c3427c35314207f620bdbb23394af50f502b1c3"} Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.200180 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4","Type":"ContainerStarted","Data":"ef34273fa04965fe17f2bbd3aa0bf4f1851554d8ea6815f2445e709987a916ff"} Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.201582 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"452c347f-4ee3-46de-ba8f-c83300966f5d","Type":"ContainerStarted","Data":"5c530e830df82888bd8da30d5e625650f7e68c8b030510bac012579aa019d8e4"} Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.202595 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29","Type":"ContainerStarted","Data":"925b22ec00321cc888cc9fbf692470a9a21f58a9e31bdea13424e9a4e8bc0f0d"} Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.203474 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2" event={"ID":"cffd57cb-2509-4a9e-8e5d-1750e4b0493e","Type":"ContainerStarted","Data":"a3b568960d9f9ff1ec39f0c731def0ddc8fa8bd09f13a349d65132b889f0a830"} Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.204574 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd","Type":"ContainerStarted","Data":"3a331050dd668115aeeac6c11892c1f6cc093d8d9d3b5acc42ae42790c6e35e0"} Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.205626 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bee11cf6-ec1b-40bb-8030-044d8f06591e","Type":"ContainerStarted","Data":"fa7415ffdccf3dfa7986ea148ea45e4b64a565d2af7811fce09af66a733bab92"} Nov 28 15:44:16 crc kubenswrapper[4805]: E1128 15:44:16.206910 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627\\\"\"" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" podUID="7c7eed7b-132f-4796-bcc7-fb991373bded" Nov 28 15:44:16 crc kubenswrapper[4805]: E1128 15:44:16.207261 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627\\\"\"" pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" podUID="86255190-b7c6-4ede-8327-196ffbceb9d0" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.645264 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.652437 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.774250 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnnhc\" (UniqueName: \"kubernetes.io/projected/b86b55ee-3e73-4509-bf86-66d380dbb8d0-kube-api-access-lnnhc\") pod \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\" (UID: \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\") " Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.774299 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86b55ee-3e73-4509-bf86-66d380dbb8d0-config\") pod \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\" (UID: \"b86b55ee-3e73-4509-bf86-66d380dbb8d0\") " Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.774477 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-dns-svc\") pod \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.774509 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-config\") pod \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.774560 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8j6l\" (UniqueName: \"kubernetes.io/projected/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-kube-api-access-n8j6l\") pod \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\" (UID: \"d0f37808-4526-4e50-ba7f-8a0974ae0bf3\") " Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.774874 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86b55ee-3e73-4509-bf86-66d380dbb8d0-config" (OuterVolumeSpecName: "config") pod "b86b55ee-3e73-4509-bf86-66d380dbb8d0" (UID: "b86b55ee-3e73-4509-bf86-66d380dbb8d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.774965 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86b55ee-3e73-4509-bf86-66d380dbb8d0-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.774967 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-config" (OuterVolumeSpecName: "config") pod "d0f37808-4526-4e50-ba7f-8a0974ae0bf3" (UID: "d0f37808-4526-4e50-ba7f-8a0974ae0bf3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.775412 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d0f37808-4526-4e50-ba7f-8a0974ae0bf3" (UID: "d0f37808-4526-4e50-ba7f-8a0974ae0bf3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.781581 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-kube-api-access-n8j6l" (OuterVolumeSpecName: "kube-api-access-n8j6l") pod "d0f37808-4526-4e50-ba7f-8a0974ae0bf3" (UID: "d0f37808-4526-4e50-ba7f-8a0974ae0bf3"). InnerVolumeSpecName "kube-api-access-n8j6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.801109 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b86b55ee-3e73-4509-bf86-66d380dbb8d0-kube-api-access-lnnhc" (OuterVolumeSpecName: "kube-api-access-lnnhc") pod "b86b55ee-3e73-4509-bf86-66d380dbb8d0" (UID: "b86b55ee-3e73-4509-bf86-66d380dbb8d0"). InnerVolumeSpecName "kube-api-access-lnnhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.876848 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.876873 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.876883 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8j6l\" (UniqueName: \"kubernetes.io/projected/d0f37808-4526-4e50-ba7f-8a0974ae0bf3-kube-api-access-n8j6l\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:16 crc kubenswrapper[4805]: I1128 15:44:16.876893 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnnhc\" (UniqueName: \"kubernetes.io/projected/b86b55ee-3e73-4509-bf86-66d380dbb8d0-kube-api-access-lnnhc\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.218078 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766fdc659c-s67c9" event={"ID":"d0f37808-4526-4e50-ba7f-8a0974ae0bf3","Type":"ContainerDied","Data":"b632c02467f203642764b049ee266848a86a725d50462b7d128911eebfc4f333"} Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.223442 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-557f57d995-xd9gt" event={"ID":"b86b55ee-3e73-4509-bf86-66d380dbb8d0","Type":"ContainerDied","Data":"9613c43646bfca34fa76accd09288ef0c0466c93cbde786e8ef93fd02196e39f"} Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.223512 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-xd9gt" Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.226102 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a","Type":"ContainerStarted","Data":"bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52"} Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.227452 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-s67c9" Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.227461 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb122aae-0e09-46b2-926c-037d25e79477","Type":"ContainerStarted","Data":"1ccc26baa6ad7cbc405abb6d95711fcbf099c77d7ab2dbe23883e5b60b452fc1"} Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.228225 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vssvf"] Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.316785 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-xd9gt"] Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.329594 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-xd9gt"] Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.348420 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-s67c9"] Nov 28 15:44:17 crc kubenswrapper[4805]: I1128 15:44:17.352320 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-s67c9"] Nov 28 15:44:19 crc kubenswrapper[4805]: I1128 15:44:19.214241 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b86b55ee-3e73-4509-bf86-66d380dbb8d0" path="/var/lib/kubelet/pods/b86b55ee-3e73-4509-bf86-66d380dbb8d0/volumes" Nov 28 15:44:19 crc kubenswrapper[4805]: I1128 15:44:19.214933 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0f37808-4526-4e50-ba7f-8a0974ae0bf3" path="/var/lib/kubelet/pods/d0f37808-4526-4e50-ba7f-8a0974ae0bf3/volumes" Nov 28 15:44:19 crc kubenswrapper[4805]: I1128 15:44:19.242514 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vssvf" event={"ID":"89d793e5-c2d1-4630-95ff-615cd30b5d04","Type":"ContainerStarted","Data":"42e8b135ecdd2e74471d5c3cd1f1a2acbffee0ec25633ec02f3b3344957ef357"} Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.275835 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vssvf" event={"ID":"89d793e5-c2d1-4630-95ff-615cd30b5d04","Type":"ContainerStarted","Data":"3752e4507479f7ee571985787e2818df7e3011388484910f0422a84a2d442fc5"} Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.292433 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29","Type":"ContainerStarted","Data":"01f4eeccdd9eae514a9fe127c422ee73a939f8ea3e4c0008f73fe1ea10f2ee26"} Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.293652 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.295297 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2" event={"ID":"cffd57cb-2509-4a9e-8e5d-1750e4b0493e","Type":"ContainerStarted","Data":"fff36b28f1a692a162fd59eb5953068bcc570d89664204a8c1cd194a602cd634"} Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.295454 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-4q9c2" Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.297381 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd","Type":"ContainerStarted","Data":"47d6ecbcb66a1126a1aefe7cd1c4f66bc52307aa387bd35f8ee20bce900a2305"} Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.299271 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bee11cf6-ec1b-40bb-8030-044d8f06591e","Type":"ContainerStarted","Data":"74986ddf8b72c947528126972ddc4887c37fae3cb83972e4872342f67b61a129"} Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.301611 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fb35eb4d-a935-4c99-a3f9-e2fc22c32157","Type":"ContainerStarted","Data":"8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508"} Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.302223 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.304167 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4","Type":"ContainerStarted","Data":"5e976116e68f4c0f2b31a953105e7f91891d03875fa4ded4f3ff8f472c7e6772"} Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.312082 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"452c347f-4ee3-46de-ba8f-c83300966f5d","Type":"ContainerStarted","Data":"834da8fff3dc97a83b4cef8f408cff6c49dd4c0dba0419975a5a61642d26191a"} Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.406051 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4q9c2" podStartSLOduration=14.284713681 podStartE2EDuration="20.406029287s" podCreationTimestamp="2025-11-28 15:44:03 +0000 UTC" firstStartedPulling="2025-11-28 15:44:16.065866565 +0000 UTC m=+1083.115657876" lastFinishedPulling="2025-11-28 15:44:22.187182151 +0000 UTC m=+1089.236973482" observedRunningTime="2025-11-28 15:44:23.402348275 +0000 UTC m=+1090.452139586" watchObservedRunningTime="2025-11-28 15:44:23.406029287 +0000 UTC m=+1090.455820608" Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.428034 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=21.368834658 podStartE2EDuration="27.428011613s" podCreationTimestamp="2025-11-28 15:43:56 +0000 UTC" firstStartedPulling="2025-11-28 15:44:15.907576905 +0000 UTC m=+1082.957368216" lastFinishedPulling="2025-11-28 15:44:21.96675386 +0000 UTC m=+1089.016545171" observedRunningTime="2025-11-28 15:44:23.423918059 +0000 UTC m=+1090.473709370" watchObservedRunningTime="2025-11-28 15:44:23.428011613 +0000 UTC m=+1090.477802934" Nov 28 15:44:23 crc kubenswrapper[4805]: I1128 15:44:23.502492 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=18.704610735 podStartE2EDuration="25.502470473s" podCreationTimestamp="2025-11-28 15:43:58 +0000 UTC" firstStartedPulling="2025-11-28 15:44:15.880741516 +0000 UTC m=+1082.930532827" lastFinishedPulling="2025-11-28 15:44:22.678601254 +0000 UTC m=+1089.728392565" observedRunningTime="2025-11-28 15:44:23.49403145 +0000 UTC m=+1090.543822761" watchObservedRunningTime="2025-11-28 15:44:23.502470473 +0000 UTC m=+1090.552261794" Nov 28 15:44:24 crc kubenswrapper[4805]: I1128 15:44:24.319851 4805 generic.go:334] "Generic (PLEG): container finished" podID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerID="3752e4507479f7ee571985787e2818df7e3011388484910f0422a84a2d442fc5" exitCode=0 Nov 28 15:44:24 crc kubenswrapper[4805]: I1128 15:44:24.321052 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vssvf" event={"ID":"89d793e5-c2d1-4630-95ff-615cd30b5d04","Type":"ContainerDied","Data":"3752e4507479f7ee571985787e2818df7e3011388484910f0422a84a2d442fc5"} Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.338336 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bee11cf6-ec1b-40bb-8030-044d8f06591e","Type":"ContainerStarted","Data":"c06570b4467d05bf08c564a535eded9242a1ad77a460690c746c25af47263da6"} Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.340515 4805 generic.go:334] "Generic (PLEG): container finished" podID="5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" containerID="5e976116e68f4c0f2b31a953105e7f91891d03875fa4ded4f3ff8f472c7e6772" exitCode=0 Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.340610 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4","Type":"ContainerDied","Data":"5e976116e68f4c0f2b31a953105e7f91891d03875fa4ded4f3ff8f472c7e6772"} Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.344536 4805 generic.go:334] "Generic (PLEG): container finished" podID="452c347f-4ee3-46de-ba8f-c83300966f5d" containerID="834da8fff3dc97a83b4cef8f408cff6c49dd4c0dba0419975a5a61642d26191a" exitCode=0 Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.344571 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"452c347f-4ee3-46de-ba8f-c83300966f5d","Type":"ContainerDied","Data":"834da8fff3dc97a83b4cef8f408cff6c49dd4c0dba0419975a5a61642d26191a"} Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.349193 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vssvf" event={"ID":"89d793e5-c2d1-4630-95ff-615cd30b5d04","Type":"ContainerStarted","Data":"e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7"} Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.349333 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vssvf" event={"ID":"89d793e5-c2d1-4630-95ff-615cd30b5d04","Type":"ContainerStarted","Data":"2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0"} Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.349408 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.351690 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd","Type":"ContainerStarted","Data":"2d959c7b97ea239b52290b670e32647011147e73ba303def59e8e383702025e0"} Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.371294 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.64637759 podStartE2EDuration="24.371273547s" podCreationTimestamp="2025-11-28 15:44:02 +0000 UTC" firstStartedPulling="2025-11-28 15:44:16.090153763 +0000 UTC m=+1083.139945074" lastFinishedPulling="2025-11-28 15:44:25.81504972 +0000 UTC m=+1092.864841031" observedRunningTime="2025-11-28 15:44:26.367524804 +0000 UTC m=+1093.417316125" watchObservedRunningTime="2025-11-28 15:44:26.371273547 +0000 UTC m=+1093.421064858" Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.390827 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.74548244 podStartE2EDuration="21.390807785s" podCreationTimestamp="2025-11-28 15:44:05 +0000 UTC" firstStartedPulling="2025-11-28 15:44:16.161148399 +0000 UTC m=+1083.210939710" lastFinishedPulling="2025-11-28 15:44:25.806473744 +0000 UTC m=+1092.856265055" observedRunningTime="2025-11-28 15:44:26.387884725 +0000 UTC m=+1093.437676056" watchObservedRunningTime="2025-11-28 15:44:26.390807785 +0000 UTC m=+1093.440599096" Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.437407 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-vssvf" podStartSLOduration=19.794613149 podStartE2EDuration="23.437389068s" podCreationTimestamp="2025-11-28 15:44:03 +0000 UTC" firstStartedPulling="2025-11-28 15:44:18.455941336 +0000 UTC m=+1085.505732647" lastFinishedPulling="2025-11-28 15:44:22.098717255 +0000 UTC m=+1089.148508566" observedRunningTime="2025-11-28 15:44:26.427417543 +0000 UTC m=+1093.477208864" watchObservedRunningTime="2025-11-28 15:44:26.437389068 +0000 UTC m=+1093.487180389" Nov 28 15:44:26 crc kubenswrapper[4805]: I1128 15:44:26.797821 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.009633 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.363144 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4","Type":"ContainerStarted","Data":"d649033916f10c5d8b5e440164330092d201d1409da92cf8c71f690663343e7c"} Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.367195 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"452c347f-4ee3-46de-ba8f-c83300966f5d","Type":"ContainerStarted","Data":"8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f"} Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.367737 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.388472 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=27.851681742 podStartE2EDuration="34.38845338s" podCreationTimestamp="2025-11-28 15:43:53 +0000 UTC" firstStartedPulling="2025-11-28 15:44:16.051646783 +0000 UTC m=+1083.101438094" lastFinishedPulling="2025-11-28 15:44:22.588418421 +0000 UTC m=+1089.638209732" observedRunningTime="2025-11-28 15:44:27.382600408 +0000 UTC m=+1094.432391729" watchObservedRunningTime="2025-11-28 15:44:27.38845338 +0000 UTC m=+1094.438244691" Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.406782 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=26.348143004 podStartE2EDuration="32.406760844s" podCreationTimestamp="2025-11-28 15:43:55 +0000 UTC" firstStartedPulling="2025-11-28 15:44:16.040116795 +0000 UTC m=+1083.089908106" lastFinishedPulling="2025-11-28 15:44:22.098734635 +0000 UTC m=+1089.148525946" observedRunningTime="2025-11-28 15:44:27.400739818 +0000 UTC m=+1094.450531159" watchObservedRunningTime="2025-11-28 15:44:27.406760844 +0000 UTC m=+1094.456552145" Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.527765 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.568110 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.798101 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:27 crc kubenswrapper[4805]: I1128 15:44:27.842727 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.379713 4805 generic.go:334] "Generic (PLEG): container finished" podID="7c7eed7b-132f-4796-bcc7-fb991373bded" containerID="125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374" exitCode=0 Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.379905 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" event={"ID":"7c7eed7b-132f-4796-bcc7-fb991373bded","Type":"ContainerDied","Data":"125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374"} Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.382928 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.426774 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.435515 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.590092 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-q6hk5"] Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.633308 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-tw6ng"] Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.634859 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.636985 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.648461 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-tw6ng"] Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.740812 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-89z6k"] Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.746788 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.749989 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.766142 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-89z6k"] Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.771740 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-ovsdbserver-nb\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.774125 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mb9s\" (UniqueName: \"kubernetes.io/projected/d3fc9922-cb39-4178-afde-284c568808ba-kube-api-access-6mb9s\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.774194 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-config\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.774289 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-dns-svc\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.810842 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b5zbf"] Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.872427 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-72ljh"] Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.873681 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875438 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59pnx\" (UniqueName: \"kubernetes.io/projected/6aff7621-5632-499d-955d-7ea5b9915b54-kube-api-access-59pnx\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875466 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aff7621-5632-499d-955d-7ea5b9915b54-config\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875496 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-combined-ca-bundle\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875524 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mb9s\" (UniqueName: \"kubernetes.io/projected/d3fc9922-cb39-4178-afde-284c568808ba-kube-api-access-6mb9s\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875545 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovs-rundir\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875560 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875580 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-config\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875605 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovn-rundir\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875631 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-dns-svc\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.875653 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-ovsdbserver-nb\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.876409 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-ovsdbserver-nb\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.877140 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-config\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.877637 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-dns-svc\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.886668 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.912847 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mb9s\" (UniqueName: \"kubernetes.io/projected/d3fc9922-cb39-4178-afde-284c568808ba-kube-api-access-6mb9s\") pod \"dnsmasq-dns-b4845d8d5-tw6ng\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.930199 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.931528 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.945513 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-72ljh"] Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.947109 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-2l2cn" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.949802 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.947180 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.947212 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.947284 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.981850 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovs-rundir\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.981888 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.981919 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovn-rundir\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.981943 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-config\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.981975 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.981993 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-dns-svc\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.982046 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59pnx\" (UniqueName: \"kubernetes.io/projected/6aff7621-5632-499d-955d-7ea5b9915b54-kube-api-access-59pnx\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.982065 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aff7621-5632-499d-955d-7ea5b9915b54-config\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.982081 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.982100 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsf4n\" (UniqueName: \"kubernetes.io/projected/743d512f-3550-4fa9-84b9-fec17810545c-kube-api-access-zsf4n\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.982137 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-combined-ca-bundle\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.983977 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovs-rundir\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.985244 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-combined-ca-bundle\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.986072 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aff7621-5632-499d-955d-7ea5b9915b54-config\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.986141 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovn-rundir\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:28 crc kubenswrapper[4805]: I1128 15:44:28.990837 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.005397 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.042626 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59pnx\" (UniqueName: \"kubernetes.io/projected/6aff7621-5632-499d-955d-7ea5b9915b54-kube-api-access-59pnx\") pod \"ovn-controller-metrics-89z6k\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.089424 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090420 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090452 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-scripts\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090478 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sfzs\" (UniqueName: \"kubernetes.io/projected/624edf72-de15-4026-812b-36d993917176-kube-api-access-2sfzs\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090493 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-config\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090541 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-config\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090560 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090587 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090602 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-dns-svc\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090629 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090675 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090694 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsf4n\" (UniqueName: \"kubernetes.io/projected/743d512f-3550-4fa9-84b9-fec17810545c-kube-api-access-zsf4n\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.090719 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/624edf72-de15-4026-812b-36d993917176-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.091415 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-config\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.091944 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.095501 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.096804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-dns-svc\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.133684 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsf4n\" (UniqueName: \"kubernetes.io/projected/743d512f-3550-4fa9-84b9-fec17810545c-kube-api-access-zsf4n\") pod \"dnsmasq-dns-5bd7c66845-72ljh\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.190980 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.196255 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.196316 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.196392 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/624edf72-de15-4026-812b-36d993917176-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.196421 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.196448 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-scripts\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.196468 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sfzs\" (UniqueName: \"kubernetes.io/projected/624edf72-de15-4026-812b-36d993917176-kube-api-access-2sfzs\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.196483 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-config\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.197180 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-config\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.197694 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.200132 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-scripts\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.201282 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/624edf72-de15-4026-812b-36d993917176-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.202114 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.204379 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.205666 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.236442 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sfzs\" (UniqueName: \"kubernetes.io/projected/624edf72-de15-4026-812b-36d993917176-kube-api-access-2sfzs\") pod \"ovn-northd-0\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.324004 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-72ljh"] Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.339684 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-4lt76"] Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.359172 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-4lt76"] Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.359321 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.380272 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.426649 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.438996 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" event={"ID":"7c7eed7b-132f-4796-bcc7-fb991373bded","Type":"ContainerStarted","Data":"7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828"} Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.439162 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" podUID="7c7eed7b-132f-4796-bcc7-fb991373bded" containerName="dnsmasq-dns" containerID="cri-o://7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828" gracePeriod=10 Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.439472 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.448295 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" event={"ID":"86255190-b7c6-4ede-8327-196ffbceb9d0","Type":"ContainerDied","Data":"8e4421410f29d8388091403fae1550f746fbfad3dadc07432d7d71155543a25c"} Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.448455 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-b5zbf" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.488620 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" podStartSLOduration=2.722324 podStartE2EDuration="37.488597266s" podCreationTimestamp="2025-11-28 15:43:52 +0000 UTC" firstStartedPulling="2025-11-28 15:43:52.92779336 +0000 UTC m=+1059.977584671" lastFinishedPulling="2025-11-28 15:44:27.694066626 +0000 UTC m=+1094.743857937" observedRunningTime="2025-11-28 15:44:29.477028658 +0000 UTC m=+1096.526819969" watchObservedRunningTime="2025-11-28 15:44:29.488597266 +0000 UTC m=+1096.538388577" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.512675 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-nb\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.512738 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-dns-svc\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.512843 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkssx\" (UniqueName: \"kubernetes.io/projected/15970fdb-4e60-4331-bd19-40db152c2303-kube-api-access-hkssx\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.512864 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-config\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.512913 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-sb\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.614045 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrv4c\" (UniqueName: \"kubernetes.io/projected/86255190-b7c6-4ede-8327-196ffbceb9d0-kube-api-access-rrv4c\") pod \"86255190-b7c6-4ede-8327-196ffbceb9d0\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.614220 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-dns-svc\") pod \"86255190-b7c6-4ede-8327-196ffbceb9d0\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.614327 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-config\") pod \"86255190-b7c6-4ede-8327-196ffbceb9d0\" (UID: \"86255190-b7c6-4ede-8327-196ffbceb9d0\") " Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.614647 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-sb\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.614698 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-nb\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.614780 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-dns-svc\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.614923 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkssx\" (UniqueName: \"kubernetes.io/projected/15970fdb-4e60-4331-bd19-40db152c2303-kube-api-access-hkssx\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.614954 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-config\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.618034 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-config\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.619567 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "86255190-b7c6-4ede-8327-196ffbceb9d0" (UID: "86255190-b7c6-4ede-8327-196ffbceb9d0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.619869 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-config" (OuterVolumeSpecName: "config") pod "86255190-b7c6-4ede-8327-196ffbceb9d0" (UID: "86255190-b7c6-4ede-8327-196ffbceb9d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.620749 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-dns-svc\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.620766 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-sb\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.621038 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-nb\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.621122 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86255190-b7c6-4ede-8327-196ffbceb9d0-kube-api-access-rrv4c" (OuterVolumeSpecName: "kube-api-access-rrv4c") pod "86255190-b7c6-4ede-8327-196ffbceb9d0" (UID: "86255190-b7c6-4ede-8327-196ffbceb9d0"). InnerVolumeSpecName "kube-api-access-rrv4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.640806 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkssx\" (UniqueName: \"kubernetes.io/projected/15970fdb-4e60-4331-bd19-40db152c2303-kube-api-access-hkssx\") pod \"dnsmasq-dns-5f6d79597f-4lt76\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.716901 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.716953 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrv4c\" (UniqueName: \"kubernetes.io/projected/86255190-b7c6-4ede-8327-196ffbceb9d0-kube-api-access-rrv4c\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.716968 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86255190-b7c6-4ede-8327-196ffbceb9d0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.755194 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.809877 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b5zbf"] Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.821636 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b5zbf"] Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.832510 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-72ljh"] Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.871902 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-89z6k"] Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.888418 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-tw6ng"] Nov 28 15:44:29 crc kubenswrapper[4805]: I1128 15:44:29.961176 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.069152 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.238335 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrzn5\" (UniqueName: \"kubernetes.io/projected/7c7eed7b-132f-4796-bcc7-fb991373bded-kube-api-access-jrzn5\") pod \"7c7eed7b-132f-4796-bcc7-fb991373bded\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.238457 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-config\") pod \"7c7eed7b-132f-4796-bcc7-fb991373bded\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.238765 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-dns-svc\") pod \"7c7eed7b-132f-4796-bcc7-fb991373bded\" (UID: \"7c7eed7b-132f-4796-bcc7-fb991373bded\") " Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.253627 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c7eed7b-132f-4796-bcc7-fb991373bded-kube-api-access-jrzn5" (OuterVolumeSpecName: "kube-api-access-jrzn5") pod "7c7eed7b-132f-4796-bcc7-fb991373bded" (UID: "7c7eed7b-132f-4796-bcc7-fb991373bded"). InnerVolumeSpecName "kube-api-access-jrzn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.277134 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-config" (OuterVolumeSpecName: "config") pod "7c7eed7b-132f-4796-bcc7-fb991373bded" (UID: "7c7eed7b-132f-4796-bcc7-fb991373bded"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.289077 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7c7eed7b-132f-4796-bcc7-fb991373bded" (UID: "7c7eed7b-132f-4796-bcc7-fb991373bded"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:30 crc kubenswrapper[4805]: W1128 15:44:30.309392 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15970fdb_4e60_4331_bd19_40db152c2303.slice/crio-2429800ddd1c05e9b28bf98cfe741ef2272cec1c76ae35135631a81fc74b7f7b WatchSource:0}: Error finding container 2429800ddd1c05e9b28bf98cfe741ef2272cec1c76ae35135631a81fc74b7f7b: Status 404 returned error can't find the container with id 2429800ddd1c05e9b28bf98cfe741ef2272cec1c76ae35135631a81fc74b7f7b Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.309637 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-4lt76"] Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.340145 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.340399 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrzn5\" (UniqueName: \"kubernetes.io/projected/7c7eed7b-132f-4796-bcc7-fb991373bded-kube-api-access-jrzn5\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.340408 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7eed7b-132f-4796-bcc7-fb991373bded-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.378157 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.379419 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c7eed7b-132f-4796-bcc7-fb991373bded" containerName="init" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.379459 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c7eed7b-132f-4796-bcc7-fb991373bded" containerName="init" Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.385618 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c7eed7b-132f-4796-bcc7-fb991373bded" containerName="dnsmasq-dns" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.385648 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c7eed7b-132f-4796-bcc7-fb991373bded" containerName="dnsmasq-dns" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.385924 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c7eed7b-132f-4796-bcc7-fb991373bded" containerName="dnsmasq-dns" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.391918 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.395385 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.395559 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.395702 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.395810 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-snrqn" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.429341 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.458878 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" event={"ID":"743d512f-3550-4fa9-84b9-fec17810545c","Type":"ContainerDied","Data":"09aadcca9d7846aa2fb8f75e13ce3e849dbe71a3e53f6ce9a57bec0e3ef89341"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.458718 4805 generic.go:334] "Generic (PLEG): container finished" podID="743d512f-3550-4fa9-84b9-fec17810545c" containerID="09aadcca9d7846aa2fb8f75e13ce3e849dbe71a3e53f6ce9a57bec0e3ef89341" exitCode=0 Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.460211 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" event={"ID":"743d512f-3550-4fa9-84b9-fec17810545c","Type":"ContainerStarted","Data":"940d75059019b7c4794f182c4dc20f46657ba68891039f007ab0e8c9bcfd588a"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.467770 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"624edf72-de15-4026-812b-36d993917176","Type":"ContainerStarted","Data":"9f01bde7a205b705624751896b1fa4f2374f5206679f5c472a61b41559d0d9b2"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.471232 4805 generic.go:334] "Generic (PLEG): container finished" podID="d3fc9922-cb39-4178-afde-284c568808ba" containerID="ad326970e0cc5144865611976d8527330b2cfbcc21c2b476c812119149abfd5d" exitCode=0 Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.471279 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" event={"ID":"d3fc9922-cb39-4178-afde-284c568808ba","Type":"ContainerDied","Data":"ad326970e0cc5144865611976d8527330b2cfbcc21c2b476c812119149abfd5d"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.471301 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" event={"ID":"d3fc9922-cb39-4178-afde-284c568808ba","Type":"ContainerStarted","Data":"4181a9441f54f219309c86f2e5de048ca4d4b875b4e721fea570491b9be037d2"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.476905 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-89z6k" event={"ID":"6aff7621-5632-499d-955d-7ea5b9915b54","Type":"ContainerStarted","Data":"0a2988e5d6f726fd2171f644ba568d97427b00e8ccb737f58ee5c66888ea03f4"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.477079 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-89z6k" event={"ID":"6aff7621-5632-499d-955d-7ea5b9915b54","Type":"ContainerStarted","Data":"950cdf1704e1f2d57fa510e2ac4b2b3f220da2b02ccee75987b07b564e026297"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.478212 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" event={"ID":"15970fdb-4e60-4331-bd19-40db152c2303","Type":"ContainerStarted","Data":"2429800ddd1c05e9b28bf98cfe741ef2272cec1c76ae35135631a81fc74b7f7b"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.480054 4805 generic.go:334] "Generic (PLEG): container finished" podID="7c7eed7b-132f-4796-bcc7-fb991373bded" containerID="7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828" exitCode=0 Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.482554 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.482584 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" event={"ID":"7c7eed7b-132f-4796-bcc7-fb991373bded","Type":"ContainerDied","Data":"7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.482627 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57dc4c6697-q6hk5" event={"ID":"7c7eed7b-132f-4796-bcc7-fb991373bded","Type":"ContainerDied","Data":"a2ece9cca3f41933f7e7a4fbd0348c83d059b5706d9f3ffe3ce31f24fd48f987"} Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.482647 4805 scope.go:117] "RemoveContainer" containerID="7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.511943 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-89z6k" podStartSLOduration=2.511923648 podStartE2EDuration="2.511923648s" podCreationTimestamp="2025-11-28 15:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:44:30.500919715 +0000 UTC m=+1097.550711046" watchObservedRunningTime="2025-11-28 15:44:30.511923648 +0000 UTC m=+1097.561714959" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.545307 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.545389 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.545414 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8scq\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-kube-api-access-j8scq\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.545447 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-cache\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.545482 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-lock\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.545954 4805 scope.go:117] "RemoveContainer" containerID="125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.598288 4805 scope.go:117] "RemoveContainer" containerID="7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828" Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.602196 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828\": container with ID starting with 7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828 not found: ID does not exist" containerID="7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.602244 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828"} err="failed to get container status \"7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828\": rpc error: code = NotFound desc = could not find container \"7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828\": container with ID starting with 7c79613444a2cb3d0a4caedabc94c007747d9e80c35206c0482d163fd5451828 not found: ID does not exist" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.602326 4805 scope.go:117] "RemoveContainer" containerID="125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374" Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.604280 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374\": container with ID starting with 125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374 not found: ID does not exist" containerID="125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.604338 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374"} err="failed to get container status \"125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374\": rpc error: code = NotFound desc = could not find container \"125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374\": container with ID starting with 125624c9b5621d99f956695287cf962b981e79f65584a212674f3b2fd3601374 not found: ID does not exist" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.610225 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-q6hk5"] Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.625501 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-q6hk5"] Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.647150 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.647206 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8scq\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-kube-api-access-j8scq\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.647283 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-cache\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.647347 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-lock\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.647622 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.648199 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.648224 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.648273 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift podName:f3d69b1a-cf80-424f-ac0c-7815e7ba0375 nodeName:}" failed. No retries permitted until 2025-11-28 15:44:31.148257002 +0000 UTC m=+1098.198048313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift") pod "swift-storage-0" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375") : configmap "swift-ring-files" not found Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.650579 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-lock\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.650822 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.651175 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-cache\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.669817 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8scq\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-kube-api-access-j8scq\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.672536 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.766945 4805 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 28 15:44:30 crc kubenswrapper[4805]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/d3fc9922-cb39-4178-afde-284c568808ba/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 28 15:44:30 crc kubenswrapper[4805]: > podSandboxID="4181a9441f54f219309c86f2e5de048ca4d4b875b4e721fea570491b9be037d2" Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.767102 4805 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 28 15:44:30 crc kubenswrapper[4805]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bfh5d7h8hd8h664h564hfbh5d4h5f5h55h5fch66h675hb8h65bh64dhbh5dchc9h66fh5dbhf4h658h64ch55bhbh65h55dh597h68dh579hbdq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6mb9s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-b4845d8d5-tw6ng_openstack(d3fc9922-cb39-4178-afde-284c568808ba): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/d3fc9922-cb39-4178-afde-284c568808ba/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 28 15:44:30 crc kubenswrapper[4805]: > logger="UnhandledError" Nov 28 15:44:30 crc kubenswrapper[4805]: E1128 15:44:30.768318 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/d3fc9922-cb39-4178-afde-284c568808ba/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" podUID="d3fc9922-cb39-4178-afde-284c568808ba" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.811426 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.953918 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-config\") pod \"743d512f-3550-4fa9-84b9-fec17810545c\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.954006 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-dns-svc\") pod \"743d512f-3550-4fa9-84b9-fec17810545c\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.954035 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-sb\") pod \"743d512f-3550-4fa9-84b9-fec17810545c\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.954072 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-nb\") pod \"743d512f-3550-4fa9-84b9-fec17810545c\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.954256 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsf4n\" (UniqueName: \"kubernetes.io/projected/743d512f-3550-4fa9-84b9-fec17810545c-kube-api-access-zsf4n\") pod \"743d512f-3550-4fa9-84b9-fec17810545c\" (UID: \"743d512f-3550-4fa9-84b9-fec17810545c\") " Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.963327 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/743d512f-3550-4fa9-84b9-fec17810545c-kube-api-access-zsf4n" (OuterVolumeSpecName: "kube-api-access-zsf4n") pod "743d512f-3550-4fa9-84b9-fec17810545c" (UID: "743d512f-3550-4fa9-84b9-fec17810545c"). InnerVolumeSpecName "kube-api-access-zsf4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.977698 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "743d512f-3550-4fa9-84b9-fec17810545c" (UID: "743d512f-3550-4fa9-84b9-fec17810545c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.979336 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "743d512f-3550-4fa9-84b9-fec17810545c" (UID: "743d512f-3550-4fa9-84b9-fec17810545c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.985002 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-config" (OuterVolumeSpecName: "config") pod "743d512f-3550-4fa9-84b9-fec17810545c" (UID: "743d512f-3550-4fa9-84b9-fec17810545c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:30 crc kubenswrapper[4805]: I1128 15:44:30.986482 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "743d512f-3550-4fa9-84b9-fec17810545c" (UID: "743d512f-3550-4fa9-84b9-fec17810545c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.056230 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsf4n\" (UniqueName: \"kubernetes.io/projected/743d512f-3550-4fa9-84b9-fec17810545c-kube-api-access-zsf4n\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.056274 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.056283 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.056291 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.056300 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/743d512f-3550-4fa9-84b9-fec17810545c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.157604 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:31 crc kubenswrapper[4805]: E1128 15:44:31.157812 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 15:44:31 crc kubenswrapper[4805]: E1128 15:44:31.157952 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 15:44:31 crc kubenswrapper[4805]: E1128 15:44:31.158004 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift podName:f3d69b1a-cf80-424f-ac0c-7815e7ba0375 nodeName:}" failed. No retries permitted until 2025-11-28 15:44:32.15798876 +0000 UTC m=+1099.207780061 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift") pod "swift-storage-0" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375") : configmap "swift-ring-files" not found Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.214765 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c7eed7b-132f-4796-bcc7-fb991373bded" path="/var/lib/kubelet/pods/7c7eed7b-132f-4796-bcc7-fb991373bded/volumes" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.215703 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86255190-b7c6-4ede-8327-196ffbceb9d0" path="/var/lib/kubelet/pods/86255190-b7c6-4ede-8327-196ffbceb9d0/volumes" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.489065 4805 generic.go:334] "Generic (PLEG): container finished" podID="15970fdb-4e60-4331-bd19-40db152c2303" containerID="8de05ca7dc13db2d90ef6ba4f2b7f9de29b8b13fb05bcb06f8eb51f955edf2a3" exitCode=0 Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.489186 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" event={"ID":"15970fdb-4e60-4331-bd19-40db152c2303","Type":"ContainerDied","Data":"8de05ca7dc13db2d90ef6ba4f2b7f9de29b8b13fb05bcb06f8eb51f955edf2a3"} Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.493901 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" event={"ID":"743d512f-3550-4fa9-84b9-fec17810545c","Type":"ContainerDied","Data":"940d75059019b7c4794f182c4dc20f46657ba68891039f007ab0e8c9bcfd588a"} Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.493945 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd7c66845-72ljh" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.493953 4805 scope.go:117] "RemoveContainer" containerID="09aadcca9d7846aa2fb8f75e13ce3e849dbe71a3e53f6ce9a57bec0e3ef89341" Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.548249 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-72ljh"] Nov 28 15:44:31 crc kubenswrapper[4805]: I1128 15:44:31.570497 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-72ljh"] Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.174620 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:32 crc kubenswrapper[4805]: E1128 15:44:32.174831 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 15:44:32 crc kubenswrapper[4805]: E1128 15:44:32.175048 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 15:44:32 crc kubenswrapper[4805]: E1128 15:44:32.175102 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift podName:f3d69b1a-cf80-424f-ac0c-7815e7ba0375 nodeName:}" failed. No retries permitted until 2025-11-28 15:44:34.17508529 +0000 UTC m=+1101.224876601 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift") pod "swift-storage-0" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375") : configmap "swift-ring-files" not found Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.503750 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" event={"ID":"15970fdb-4e60-4331-bd19-40db152c2303","Type":"ContainerStarted","Data":"05de3845991eaa359e6baa626dc02171b7f292526b958ada1028d01c0e1669d6"} Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.504908 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.509800 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"624edf72-de15-4026-812b-36d993917176","Type":"ContainerStarted","Data":"fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574"} Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.510091 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.510102 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"624edf72-de15-4026-812b-36d993917176","Type":"ContainerStarted","Data":"10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d"} Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.512345 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" event={"ID":"d3fc9922-cb39-4178-afde-284c568808ba","Type":"ContainerStarted","Data":"cb7ba7f6c2aec3276cb7b4139093d3b82483611c6c13f1ee5c0892b22597e31d"} Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.512571 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.525776 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" podStartSLOduration=3.525758077 podStartE2EDuration="3.525758077s" podCreationTimestamp="2025-11-28 15:44:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:44:32.520713898 +0000 UTC m=+1099.570505209" watchObservedRunningTime="2025-11-28 15:44:32.525758077 +0000 UTC m=+1099.575549388" Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.541824 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" podStartSLOduration=4.541796709 podStartE2EDuration="4.541796709s" podCreationTimestamp="2025-11-28 15:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:44:32.536611846 +0000 UTC m=+1099.586403157" watchObservedRunningTime="2025-11-28 15:44:32.541796709 +0000 UTC m=+1099.591588060" Nov 28 15:44:32 crc kubenswrapper[4805]: I1128 15:44:32.566941 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.192540221 podStartE2EDuration="4.566918811s" podCreationTimestamp="2025-11-28 15:44:28 +0000 UTC" firstStartedPulling="2025-11-28 15:44:29.984730769 +0000 UTC m=+1097.034522080" lastFinishedPulling="2025-11-28 15:44:31.359109359 +0000 UTC m=+1098.408900670" observedRunningTime="2025-11-28 15:44:32.560673719 +0000 UTC m=+1099.610465030" watchObservedRunningTime="2025-11-28 15:44:32.566918811 +0000 UTC m=+1099.616710122" Nov 28 15:44:33 crc kubenswrapper[4805]: I1128 15:44:33.215583 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="743d512f-3550-4fa9-84b9-fec17810545c" path="/var/lib/kubelet/pods/743d512f-3550-4fa9-84b9-fec17810545c/volumes" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.214094 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:34 crc kubenswrapper[4805]: E1128 15:44:34.214436 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 15:44:34 crc kubenswrapper[4805]: E1128 15:44:34.214479 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 15:44:34 crc kubenswrapper[4805]: E1128 15:44:34.214584 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift podName:f3d69b1a-cf80-424f-ac0c-7815e7ba0375 nodeName:}" failed. No retries permitted until 2025-11-28 15:44:38.214555055 +0000 UTC m=+1105.264346396 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift") pod "swift-storage-0" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375") : configmap "swift-ring-files" not found Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.257893 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-7f2hz"] Nov 28 15:44:34 crc kubenswrapper[4805]: E1128 15:44:34.258701 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743d512f-3550-4fa9-84b9-fec17810545c" containerName="init" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.258752 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="743d512f-3550-4fa9-84b9-fec17810545c" containerName="init" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.259230 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="743d512f-3550-4fa9-84b9-fec17810545c" containerName="init" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.260509 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.267933 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.268541 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.273056 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7f2hz"] Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.273227 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.315823 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-ring-data-devices\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.315903 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgw28\" (UniqueName: \"kubernetes.io/projected/300c24d4-04ee-4913-8e98-954c3f5c6f94-kube-api-access-qgw28\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.315948 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-scripts\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.316004 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-combined-ca-bundle\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.316147 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-swiftconf\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.316177 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/300c24d4-04ee-4913-8e98-954c3f5c6f94-etc-swift\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.316205 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-dispersionconf\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.417648 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-ring-data-devices\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.417695 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgw28\" (UniqueName: \"kubernetes.io/projected/300c24d4-04ee-4913-8e98-954c3f5c6f94-kube-api-access-qgw28\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.417726 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-scripts\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.417779 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-combined-ca-bundle\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.417825 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-swiftconf\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.417848 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/300c24d4-04ee-4913-8e98-954c3f5c6f94-etc-swift\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.417875 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-dispersionconf\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.419115 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-ring-data-devices\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.419253 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/300c24d4-04ee-4913-8e98-954c3f5c6f94-etc-swift\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.420745 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-scripts\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.427515 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-combined-ca-bundle\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.428240 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-swiftconf\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.428853 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-dispersionconf\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.450276 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgw28\" (UniqueName: \"kubernetes.io/projected/300c24d4-04ee-4913-8e98-954c3f5c6f94-kube-api-access-qgw28\") pod \"swift-ring-rebalance-7f2hz\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:34 crc kubenswrapper[4805]: I1128 15:44:34.594191 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:35 crc kubenswrapper[4805]: I1128 15:44:35.137508 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7f2hz"] Nov 28 15:44:35 crc kubenswrapper[4805]: W1128 15:44:35.155720 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod300c24d4_04ee_4913_8e98_954c3f5c6f94.slice/crio-fcc8807240cd9ee3bfb821e919b131cb12251b53ed44a7e4e6da512446fff68e WatchSource:0}: Error finding container fcc8807240cd9ee3bfb821e919b131cb12251b53ed44a7e4e6da512446fff68e: Status 404 returned error can't find the container with id fcc8807240cd9ee3bfb821e919b131cb12251b53ed44a7e4e6da512446fff68e Nov 28 15:44:35 crc kubenswrapper[4805]: I1128 15:44:35.538418 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7f2hz" event={"ID":"300c24d4-04ee-4913-8e98-954c3f5c6f94","Type":"ContainerStarted","Data":"fcc8807240cd9ee3bfb821e919b131cb12251b53ed44a7e4e6da512446fff68e"} Nov 28 15:44:35 crc kubenswrapper[4805]: I1128 15:44:35.608507 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 28 15:44:35 crc kubenswrapper[4805]: I1128 15:44:35.608565 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 28 15:44:35 crc kubenswrapper[4805]: I1128 15:44:35.723500 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.632909 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.654571 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.655404 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.765758 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.925809 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bd4b-account-create-update-zmxqb"] Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.926785 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.934437 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bd4b-account-create-update-zmxqb"] Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.934741 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.975882 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc5h5\" (UniqueName: \"kubernetes.io/projected/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-kube-api-access-xc5h5\") pod \"placement-bd4b-account-create-update-zmxqb\" (UID: \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\") " pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.975943 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-operator-scripts\") pod \"placement-bd4b-account-create-update-zmxqb\" (UID: \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\") " pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.993404 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-ph6gw"] Nov 28 15:44:36 crc kubenswrapper[4805]: I1128 15:44:36.994883 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.002633 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ph6gw"] Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.082960 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc5h5\" (UniqueName: \"kubernetes.io/projected/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-kube-api-access-xc5h5\") pod \"placement-bd4b-account-create-update-zmxqb\" (UID: \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\") " pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.083286 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnmgs\" (UniqueName: \"kubernetes.io/projected/e937c330-5ffb-48ee-8607-fa0764ef7172-kube-api-access-nnmgs\") pod \"placement-db-create-ph6gw\" (UID: \"e937c330-5ffb-48ee-8607-fa0764ef7172\") " pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.083315 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e937c330-5ffb-48ee-8607-fa0764ef7172-operator-scripts\") pod \"placement-db-create-ph6gw\" (UID: \"e937c330-5ffb-48ee-8607-fa0764ef7172\") " pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.083336 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-operator-scripts\") pod \"placement-bd4b-account-create-update-zmxqb\" (UID: \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\") " pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.084668 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-operator-scripts\") pod \"placement-bd4b-account-create-update-zmxqb\" (UID: \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\") " pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.106681 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc5h5\" (UniqueName: \"kubernetes.io/projected/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-kube-api-access-xc5h5\") pod \"placement-bd4b-account-create-update-zmxqb\" (UID: \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\") " pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.182108 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-f5hv2"] Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.183189 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.185094 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnmgs\" (UniqueName: \"kubernetes.io/projected/e937c330-5ffb-48ee-8607-fa0764ef7172-kube-api-access-nnmgs\") pod \"placement-db-create-ph6gw\" (UID: \"e937c330-5ffb-48ee-8607-fa0764ef7172\") " pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.185203 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e937c330-5ffb-48ee-8607-fa0764ef7172-operator-scripts\") pod \"placement-db-create-ph6gw\" (UID: \"e937c330-5ffb-48ee-8607-fa0764ef7172\") " pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.191667 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e937c330-5ffb-48ee-8607-fa0764ef7172-operator-scripts\") pod \"placement-db-create-ph6gw\" (UID: \"e937c330-5ffb-48ee-8607-fa0764ef7172\") " pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.201606 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnmgs\" (UniqueName: \"kubernetes.io/projected/e937c330-5ffb-48ee-8607-fa0764ef7172-kube-api-access-nnmgs\") pod \"placement-db-create-ph6gw\" (UID: \"e937c330-5ffb-48ee-8607-fa0764ef7172\") " pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.214915 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-f5hv2"] Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.243441 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.286415 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljr7m\" (UniqueName: \"kubernetes.io/projected/89d423d2-d06f-4c64-9312-2258d0a746a6-kube-api-access-ljr7m\") pod \"glance-db-create-f5hv2\" (UID: \"89d423d2-d06f-4c64-9312-2258d0a746a6\") " pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.286769 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89d423d2-d06f-4c64-9312-2258d0a746a6-operator-scripts\") pod \"glance-db-create-f5hv2\" (UID: \"89d423d2-d06f-4c64-9312-2258d0a746a6\") " pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.289713 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-73a6-account-create-update-nfdpv"] Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.291045 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.294924 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.309844 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.311034 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-73a6-account-create-update-nfdpv"] Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.388276 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljr7m\" (UniqueName: \"kubernetes.io/projected/89d423d2-d06f-4c64-9312-2258d0a746a6-kube-api-access-ljr7m\") pod \"glance-db-create-f5hv2\" (UID: \"89d423d2-d06f-4c64-9312-2258d0a746a6\") " pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.388396 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89d423d2-d06f-4c64-9312-2258d0a746a6-operator-scripts\") pod \"glance-db-create-f5hv2\" (UID: \"89d423d2-d06f-4c64-9312-2258d0a746a6\") " pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.388516 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ff5237a-ad35-4875-9cd7-046cfbef7506-operator-scripts\") pod \"glance-73a6-account-create-update-nfdpv\" (UID: \"4ff5237a-ad35-4875-9cd7-046cfbef7506\") " pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.388542 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hr8b\" (UniqueName: \"kubernetes.io/projected/4ff5237a-ad35-4875-9cd7-046cfbef7506-kube-api-access-8hr8b\") pod \"glance-73a6-account-create-update-nfdpv\" (UID: \"4ff5237a-ad35-4875-9cd7-046cfbef7506\") " pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.389571 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89d423d2-d06f-4c64-9312-2258d0a746a6-operator-scripts\") pod \"glance-db-create-f5hv2\" (UID: \"89d423d2-d06f-4c64-9312-2258d0a746a6\") " pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.403553 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljr7m\" (UniqueName: \"kubernetes.io/projected/89d423d2-d06f-4c64-9312-2258d0a746a6-kube-api-access-ljr7m\") pod \"glance-db-create-f5hv2\" (UID: \"89d423d2-d06f-4c64-9312-2258d0a746a6\") " pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.490735 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ff5237a-ad35-4875-9cd7-046cfbef7506-operator-scripts\") pod \"glance-73a6-account-create-update-nfdpv\" (UID: \"4ff5237a-ad35-4875-9cd7-046cfbef7506\") " pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.490780 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hr8b\" (UniqueName: \"kubernetes.io/projected/4ff5237a-ad35-4875-9cd7-046cfbef7506-kube-api-access-8hr8b\") pod \"glance-73a6-account-create-update-nfdpv\" (UID: \"4ff5237a-ad35-4875-9cd7-046cfbef7506\") " pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.491607 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ff5237a-ad35-4875-9cd7-046cfbef7506-operator-scripts\") pod \"glance-73a6-account-create-update-nfdpv\" (UID: \"4ff5237a-ad35-4875-9cd7-046cfbef7506\") " pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.506085 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hr8b\" (UniqueName: \"kubernetes.io/projected/4ff5237a-ad35-4875-9cd7-046cfbef7506-kube-api-access-8hr8b\") pod \"glance-73a6-account-create-update-nfdpv\" (UID: \"4ff5237a-ad35-4875-9cd7-046cfbef7506\") " pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.545814 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.614186 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:37 crc kubenswrapper[4805]: I1128 15:44:37.647277 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 28 15:44:38 crc kubenswrapper[4805]: I1128 15:44:38.304272 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:38 crc kubenswrapper[4805]: E1128 15:44:38.304484 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 15:44:38 crc kubenswrapper[4805]: E1128 15:44:38.304957 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 15:44:38 crc kubenswrapper[4805]: E1128 15:44:38.305044 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift podName:f3d69b1a-cf80-424f-ac0c-7815e7ba0375 nodeName:}" failed. No retries permitted until 2025-11-28 15:44:46.305019516 +0000 UTC m=+1113.354810847 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift") pod "swift-storage-0" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375") : configmap "swift-ring-files" not found Nov 28 15:44:38 crc kubenswrapper[4805]: I1128 15:44:38.951534 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:39 crc kubenswrapper[4805]: I1128 15:44:39.756542 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:44:39 crc kubenswrapper[4805]: I1128 15:44:39.815990 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-tw6ng"] Nov 28 15:44:39 crc kubenswrapper[4805]: I1128 15:44:39.816278 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" podUID="d3fc9922-cb39-4178-afde-284c568808ba" containerName="dnsmasq-dns" containerID="cri-o://cb7ba7f6c2aec3276cb7b4139093d3b82483611c6c13f1ee5c0892b22597e31d" gracePeriod=10 Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.630270 4805 generic.go:334] "Generic (PLEG): container finished" podID="d3fc9922-cb39-4178-afde-284c568808ba" containerID="cb7ba7f6c2aec3276cb7b4139093d3b82483611c6c13f1ee5c0892b22597e31d" exitCode=0 Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.631420 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" event={"ID":"d3fc9922-cb39-4178-afde-284c568808ba","Type":"ContainerDied","Data":"cb7ba7f6c2aec3276cb7b4139093d3b82483611c6c13f1ee5c0892b22597e31d"} Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.757579 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.844458 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-config\") pod \"d3fc9922-cb39-4178-afde-284c568808ba\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.844511 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mb9s\" (UniqueName: \"kubernetes.io/projected/d3fc9922-cb39-4178-afde-284c568808ba-kube-api-access-6mb9s\") pod \"d3fc9922-cb39-4178-afde-284c568808ba\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.844545 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-dns-svc\") pod \"d3fc9922-cb39-4178-afde-284c568808ba\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.844587 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-ovsdbserver-nb\") pod \"d3fc9922-cb39-4178-afde-284c568808ba\" (UID: \"d3fc9922-cb39-4178-afde-284c568808ba\") " Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.853581 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3fc9922-cb39-4178-afde-284c568808ba-kube-api-access-6mb9s" (OuterVolumeSpecName: "kube-api-access-6mb9s") pod "d3fc9922-cb39-4178-afde-284c568808ba" (UID: "d3fc9922-cb39-4178-afde-284c568808ba"). InnerVolumeSpecName "kube-api-access-6mb9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.886056 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-config" (OuterVolumeSpecName: "config") pod "d3fc9922-cb39-4178-afde-284c568808ba" (UID: "d3fc9922-cb39-4178-afde-284c568808ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.889122 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d3fc9922-cb39-4178-afde-284c568808ba" (UID: "d3fc9922-cb39-4178-afde-284c568808ba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.894343 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3fc9922-cb39-4178-afde-284c568808ba" (UID: "d3fc9922-cb39-4178-afde-284c568808ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.946739 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.946780 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mb9s\" (UniqueName: \"kubernetes.io/projected/d3fc9922-cb39-4178-afde-284c568808ba-kube-api-access-6mb9s\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.946795 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.946806 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3fc9922-cb39-4178-afde-284c568808ba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:40 crc kubenswrapper[4805]: I1128 15:44:40.990087 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bd4b-account-create-update-zmxqb"] Nov 28 15:44:40 crc kubenswrapper[4805]: W1128 15:44:40.995785 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c22dd95_b8ab_44d5_8783_73ece53f5dc8.slice/crio-67e1a514d6d1788a4eb63337930070dd0b7cb633cc9c4eaec8fb4477adbcfbed WatchSource:0}: Error finding container 67e1a514d6d1788a4eb63337930070dd0b7cb633cc9c4eaec8fb4477adbcfbed: Status 404 returned error can't find the container with id 67e1a514d6d1788a4eb63337930070dd0b7cb633cc9c4eaec8fb4477adbcfbed Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.059797 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.059850 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:44:41 crc kubenswrapper[4805]: W1128 15:44:41.074746 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode937c330_5ffb_48ee_8607_fa0764ef7172.slice/crio-f1dedd608b03abed68edd95eff4a9c9b29635e71d4bb45be0b80557904e1ad09 WatchSource:0}: Error finding container f1dedd608b03abed68edd95eff4a9c9b29635e71d4bb45be0b80557904e1ad09: Status 404 returned error can't find the container with id f1dedd608b03abed68edd95eff4a9c9b29635e71d4bb45be0b80557904e1ad09 Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.076434 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ph6gw"] Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.106181 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-f5hv2"] Nov 28 15:44:41 crc kubenswrapper[4805]: W1128 15:44:41.117486 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89d423d2_d06f_4c64_9312_2258d0a746a6.slice/crio-79d4157fcae8b56e981d70a9dbd4beb209b2a4536d31eb1f0347106be8405d92 WatchSource:0}: Error finding container 79d4157fcae8b56e981d70a9dbd4beb209b2a4536d31eb1f0347106be8405d92: Status 404 returned error can't find the container with id 79d4157fcae8b56e981d70a9dbd4beb209b2a4536d31eb1f0347106be8405d92 Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.131570 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-73a6-account-create-update-nfdpv"] Nov 28 15:44:41 crc kubenswrapper[4805]: W1128 15:44:41.167189 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ff5237a_ad35_4875_9cd7_046cfbef7506.slice/crio-0385648c95d8bd9ef5d9c0123fffc26ce78af02f443685063127554d835fc8d7 WatchSource:0}: Error finding container 0385648c95d8bd9ef5d9c0123fffc26ce78af02f443685063127554d835fc8d7: Status 404 returned error can't find the container with id 0385648c95d8bd9ef5d9c0123fffc26ce78af02f443685063127554d835fc8d7 Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.638714 4805 generic.go:334] "Generic (PLEG): container finished" podID="0c22dd95-b8ab-44d5-8783-73ece53f5dc8" containerID="25f906e361ec4998574fde2ff49ffc6709f4fa6fb833757df62d571312c0b805" exitCode=0 Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.638802 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd4b-account-create-update-zmxqb" event={"ID":"0c22dd95-b8ab-44d5-8783-73ece53f5dc8","Type":"ContainerDied","Data":"25f906e361ec4998574fde2ff49ffc6709f4fa6fb833757df62d571312c0b805"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.638835 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd4b-account-create-update-zmxqb" event={"ID":"0c22dd95-b8ab-44d5-8783-73ece53f5dc8","Type":"ContainerStarted","Data":"67e1a514d6d1788a4eb63337930070dd0b7cb633cc9c4eaec8fb4477adbcfbed"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.641755 4805 generic.go:334] "Generic (PLEG): container finished" podID="4ff5237a-ad35-4875-9cd7-046cfbef7506" containerID="cc6c21d34e2ac7e3ae4c9f58a0944af49da732e4fdc04209652d7d7e9cd2a44e" exitCode=0 Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.641835 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-73a6-account-create-update-nfdpv" event={"ID":"4ff5237a-ad35-4875-9cd7-046cfbef7506","Type":"ContainerDied","Data":"cc6c21d34e2ac7e3ae4c9f58a0944af49da732e4fdc04209652d7d7e9cd2a44e"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.641868 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-73a6-account-create-update-nfdpv" event={"ID":"4ff5237a-ad35-4875-9cd7-046cfbef7506","Type":"ContainerStarted","Data":"0385648c95d8bd9ef5d9c0123fffc26ce78af02f443685063127554d835fc8d7"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.644276 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" event={"ID":"d3fc9922-cb39-4178-afde-284c568808ba","Type":"ContainerDied","Data":"4181a9441f54f219309c86f2e5de048ca4d4b875b4e721fea570491b9be037d2"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.644289 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4845d8d5-tw6ng" Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.644324 4805 scope.go:117] "RemoveContainer" containerID="cb7ba7f6c2aec3276cb7b4139093d3b82483611c6c13f1ee5c0892b22597e31d" Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.647137 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7f2hz" event={"ID":"300c24d4-04ee-4913-8e98-954c3f5c6f94","Type":"ContainerStarted","Data":"65d607a09a90f0f414b90b8d0bc2013bb6501ea2d7d313a6ba0262c0e54abf2d"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.652877 4805 generic.go:334] "Generic (PLEG): container finished" podID="89d423d2-d06f-4c64-9312-2258d0a746a6" containerID="ac30b0deb20776f06cef1ad68dfa668fd58449a1db45c10823b51c020d31976c" exitCode=0 Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.652959 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f5hv2" event={"ID":"89d423d2-d06f-4c64-9312-2258d0a746a6","Type":"ContainerDied","Data":"ac30b0deb20776f06cef1ad68dfa668fd58449a1db45c10823b51c020d31976c"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.652989 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f5hv2" event={"ID":"89d423d2-d06f-4c64-9312-2258d0a746a6","Type":"ContainerStarted","Data":"79d4157fcae8b56e981d70a9dbd4beb209b2a4536d31eb1f0347106be8405d92"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.680965 4805 generic.go:334] "Generic (PLEG): container finished" podID="e937c330-5ffb-48ee-8607-fa0764ef7172" containerID="32d65daedeab95d64fa9e1a47825872f67c3ce768660a116428eaf18af16ec3f" exitCode=0 Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.681056 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ph6gw" event={"ID":"e937c330-5ffb-48ee-8607-fa0764ef7172","Type":"ContainerDied","Data":"32d65daedeab95d64fa9e1a47825872f67c3ce768660a116428eaf18af16ec3f"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.681090 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ph6gw" event={"ID":"e937c330-5ffb-48ee-8607-fa0764ef7172","Type":"ContainerStarted","Data":"f1dedd608b03abed68edd95eff4a9c9b29635e71d4bb45be0b80557904e1ad09"} Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.695323 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-7f2hz" podStartSLOduration=2.338120653 podStartE2EDuration="7.695281604s" podCreationTimestamp="2025-11-28 15:44:34 +0000 UTC" firstStartedPulling="2025-11-28 15:44:35.161971056 +0000 UTC m=+1102.211762367" lastFinishedPulling="2025-11-28 15:44:40.519132007 +0000 UTC m=+1107.568923318" observedRunningTime="2025-11-28 15:44:41.689891977 +0000 UTC m=+1108.739683298" watchObservedRunningTime="2025-11-28 15:44:41.695281604 +0000 UTC m=+1108.745072925" Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.711003 4805 scope.go:117] "RemoveContainer" containerID="ad326970e0cc5144865611976d8527330b2cfbcc21c2b476c812119149abfd5d" Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.743500 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-tw6ng"] Nov 28 15:44:41 crc kubenswrapper[4805]: I1128 15:44:41.751569 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-tw6ng"] Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.131893 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.198879 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnmgs\" (UniqueName: \"kubernetes.io/projected/e937c330-5ffb-48ee-8607-fa0764ef7172-kube-api-access-nnmgs\") pod \"e937c330-5ffb-48ee-8607-fa0764ef7172\" (UID: \"e937c330-5ffb-48ee-8607-fa0764ef7172\") " Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.198943 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e937c330-5ffb-48ee-8607-fa0764ef7172-operator-scripts\") pod \"e937c330-5ffb-48ee-8607-fa0764ef7172\" (UID: \"e937c330-5ffb-48ee-8607-fa0764ef7172\") " Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.199742 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e937c330-5ffb-48ee-8607-fa0764ef7172-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e937c330-5ffb-48ee-8607-fa0764ef7172" (UID: "e937c330-5ffb-48ee-8607-fa0764ef7172"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.205884 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e937c330-5ffb-48ee-8607-fa0764ef7172-kube-api-access-nnmgs" (OuterVolumeSpecName: "kube-api-access-nnmgs") pod "e937c330-5ffb-48ee-8607-fa0764ef7172" (UID: "e937c330-5ffb-48ee-8607-fa0764ef7172"). InnerVolumeSpecName "kube-api-access-nnmgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.229269 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3fc9922-cb39-4178-afde-284c568808ba" path="/var/lib/kubelet/pods/d3fc9922-cb39-4178-afde-284c568808ba/volumes" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.254661 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.274406 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.300207 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljr7m\" (UniqueName: \"kubernetes.io/projected/89d423d2-d06f-4c64-9312-2258d0a746a6-kube-api-access-ljr7m\") pod \"89d423d2-d06f-4c64-9312-2258d0a746a6\" (UID: \"89d423d2-d06f-4c64-9312-2258d0a746a6\") " Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.300255 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89d423d2-d06f-4c64-9312-2258d0a746a6-operator-scripts\") pod \"89d423d2-d06f-4c64-9312-2258d0a746a6\" (UID: \"89d423d2-d06f-4c64-9312-2258d0a746a6\") " Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.300340 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hr8b\" (UniqueName: \"kubernetes.io/projected/4ff5237a-ad35-4875-9cd7-046cfbef7506-kube-api-access-8hr8b\") pod \"4ff5237a-ad35-4875-9cd7-046cfbef7506\" (UID: \"4ff5237a-ad35-4875-9cd7-046cfbef7506\") " Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.301534 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ff5237a-ad35-4875-9cd7-046cfbef7506-operator-scripts\") pod \"4ff5237a-ad35-4875-9cd7-046cfbef7506\" (UID: \"4ff5237a-ad35-4875-9cd7-046cfbef7506\") " Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.302073 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnmgs\" (UniqueName: \"kubernetes.io/projected/e937c330-5ffb-48ee-8607-fa0764ef7172-kube-api-access-nnmgs\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.302099 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e937c330-5ffb-48ee-8607-fa0764ef7172-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.304688 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.304985 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89d423d2-d06f-4c64-9312-2258d0a746a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "89d423d2-d06f-4c64-9312-2258d0a746a6" (UID: "89d423d2-d06f-4c64-9312-2258d0a746a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.304986 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff5237a-ad35-4875-9cd7-046cfbef7506-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ff5237a-ad35-4875-9cd7-046cfbef7506" (UID: "4ff5237a-ad35-4875-9cd7-046cfbef7506"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.307862 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ff5237a-ad35-4875-9cd7-046cfbef7506-kube-api-access-8hr8b" (OuterVolumeSpecName: "kube-api-access-8hr8b") pod "4ff5237a-ad35-4875-9cd7-046cfbef7506" (UID: "4ff5237a-ad35-4875-9cd7-046cfbef7506"). InnerVolumeSpecName "kube-api-access-8hr8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.309792 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89d423d2-d06f-4c64-9312-2258d0a746a6-kube-api-access-ljr7m" (OuterVolumeSpecName: "kube-api-access-ljr7m") pod "89d423d2-d06f-4c64-9312-2258d0a746a6" (UID: "89d423d2-d06f-4c64-9312-2258d0a746a6"). InnerVolumeSpecName "kube-api-access-ljr7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.402819 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-operator-scripts\") pod \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\" (UID: \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\") " Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.402953 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc5h5\" (UniqueName: \"kubernetes.io/projected/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-kube-api-access-xc5h5\") pod \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\" (UID: \"0c22dd95-b8ab-44d5-8783-73ece53f5dc8\") " Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.403337 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c22dd95-b8ab-44d5-8783-73ece53f5dc8" (UID: "0c22dd95-b8ab-44d5-8783-73ece53f5dc8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.403847 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljr7m\" (UniqueName: \"kubernetes.io/projected/89d423d2-d06f-4c64-9312-2258d0a746a6-kube-api-access-ljr7m\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.403866 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89d423d2-d06f-4c64-9312-2258d0a746a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.403876 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hr8b\" (UniqueName: \"kubernetes.io/projected/4ff5237a-ad35-4875-9cd7-046cfbef7506-kube-api-access-8hr8b\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.403887 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ff5237a-ad35-4875-9cd7-046cfbef7506-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.403897 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.405641 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-kube-api-access-xc5h5" (OuterVolumeSpecName: "kube-api-access-xc5h5") pod "0c22dd95-b8ab-44d5-8783-73ece53f5dc8" (UID: "0c22dd95-b8ab-44d5-8783-73ece53f5dc8"). InnerVolumeSpecName "kube-api-access-xc5h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.505468 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc5h5\" (UniqueName: \"kubernetes.io/projected/0c22dd95-b8ab-44d5-8783-73ece53f5dc8-kube-api-access-xc5h5\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.710138 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-73a6-account-create-update-nfdpv" event={"ID":"4ff5237a-ad35-4875-9cd7-046cfbef7506","Type":"ContainerDied","Data":"0385648c95d8bd9ef5d9c0123fffc26ce78af02f443685063127554d835fc8d7"} Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.710195 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0385648c95d8bd9ef5d9c0123fffc26ce78af02f443685063127554d835fc8d7" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.710270 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-73a6-account-create-update-nfdpv" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.718908 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f5hv2" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.718949 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f5hv2" event={"ID":"89d423d2-d06f-4c64-9312-2258d0a746a6","Type":"ContainerDied","Data":"79d4157fcae8b56e981d70a9dbd4beb209b2a4536d31eb1f0347106be8405d92"} Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.719014 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79d4157fcae8b56e981d70a9dbd4beb209b2a4536d31eb1f0347106be8405d92" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.721693 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ph6gw" event={"ID":"e937c330-5ffb-48ee-8607-fa0764ef7172","Type":"ContainerDied","Data":"f1dedd608b03abed68edd95eff4a9c9b29635e71d4bb45be0b80557904e1ad09"} Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.721746 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1dedd608b03abed68edd95eff4a9c9b29635e71d4bb45be0b80557904e1ad09" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.721718 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ph6gw" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.723660 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd4b-account-create-update-zmxqb" event={"ID":"0c22dd95-b8ab-44d5-8783-73ece53f5dc8","Type":"ContainerDied","Data":"67e1a514d6d1788a4eb63337930070dd0b7cb633cc9c4eaec8fb4477adbcfbed"} Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.723686 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67e1a514d6d1788a4eb63337930070dd0b7cb633cc9c4eaec8fb4477adbcfbed" Nov 28 15:44:43 crc kubenswrapper[4805]: I1128 15:44:43.723725 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd4b-account-create-update-zmxqb" Nov 28 15:44:44 crc kubenswrapper[4805]: I1128 15:44:44.435861 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.363011 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:44:46 crc kubenswrapper[4805]: E1128 15:44:46.363206 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 15:44:46 crc kubenswrapper[4805]: E1128 15:44:46.363583 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 15:44:46 crc kubenswrapper[4805]: E1128 15:44:46.363637 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift podName:f3d69b1a-cf80-424f-ac0c-7815e7ba0375 nodeName:}" failed. No retries permitted until 2025-11-28 15:45:02.363619472 +0000 UTC m=+1129.413410783 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift") pod "swift-storage-0" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375") : configmap "swift-ring-files" not found Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.611401 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-vf6sz"] Nov 28 15:44:46 crc kubenswrapper[4805]: E1128 15:44:46.611930 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d423d2-d06f-4c64-9312-2258d0a746a6" containerName="mariadb-database-create" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.611968 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d423d2-d06f-4c64-9312-2258d0a746a6" containerName="mariadb-database-create" Nov 28 15:44:46 crc kubenswrapper[4805]: E1128 15:44:46.612007 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e937c330-5ffb-48ee-8607-fa0764ef7172" containerName="mariadb-database-create" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612021 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e937c330-5ffb-48ee-8607-fa0764ef7172" containerName="mariadb-database-create" Nov 28 15:44:46 crc kubenswrapper[4805]: E1128 15:44:46.612053 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fc9922-cb39-4178-afde-284c568808ba" containerName="dnsmasq-dns" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612068 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fc9922-cb39-4178-afde-284c568808ba" containerName="dnsmasq-dns" Nov 28 15:44:46 crc kubenswrapper[4805]: E1128 15:44:46.612094 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff5237a-ad35-4875-9cd7-046cfbef7506" containerName="mariadb-account-create-update" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612108 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff5237a-ad35-4875-9cd7-046cfbef7506" containerName="mariadb-account-create-update" Nov 28 15:44:46 crc kubenswrapper[4805]: E1128 15:44:46.612132 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fc9922-cb39-4178-afde-284c568808ba" containerName="init" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612145 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fc9922-cb39-4178-afde-284c568808ba" containerName="init" Nov 28 15:44:46 crc kubenswrapper[4805]: E1128 15:44:46.612168 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c22dd95-b8ab-44d5-8783-73ece53f5dc8" containerName="mariadb-account-create-update" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612187 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c22dd95-b8ab-44d5-8783-73ece53f5dc8" containerName="mariadb-account-create-update" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612555 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e937c330-5ffb-48ee-8607-fa0764ef7172" containerName="mariadb-database-create" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612601 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="89d423d2-d06f-4c64-9312-2258d0a746a6" containerName="mariadb-database-create" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612629 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fc9922-cb39-4178-afde-284c568808ba" containerName="dnsmasq-dns" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612652 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c22dd95-b8ab-44d5-8783-73ece53f5dc8" containerName="mariadb-account-create-update" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.612675 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ff5237a-ad35-4875-9cd7-046cfbef7506" containerName="mariadb-account-create-update" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.613612 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.630027 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vf6sz"] Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.668650 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc4053dd-426b-473b-9ac2-039a55ab62a7-operator-scripts\") pod \"keystone-db-create-vf6sz\" (UID: \"bc4053dd-426b-473b-9ac2-039a55ab62a7\") " pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.668710 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz9gw\" (UniqueName: \"kubernetes.io/projected/bc4053dd-426b-473b-9ac2-039a55ab62a7-kube-api-access-sz9gw\") pod \"keystone-db-create-vf6sz\" (UID: \"bc4053dd-426b-473b-9ac2-039a55ab62a7\") " pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.714101 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-fa91-account-create-update-k9xs4"] Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.717420 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.722265 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.732380 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fa91-account-create-update-k9xs4"] Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.770257 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc4053dd-426b-473b-9ac2-039a55ab62a7-operator-scripts\") pod \"keystone-db-create-vf6sz\" (UID: \"bc4053dd-426b-473b-9ac2-039a55ab62a7\") " pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.770311 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz9gw\" (UniqueName: \"kubernetes.io/projected/bc4053dd-426b-473b-9ac2-039a55ab62a7-kube-api-access-sz9gw\") pod \"keystone-db-create-vf6sz\" (UID: \"bc4053dd-426b-473b-9ac2-039a55ab62a7\") " pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.770382 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4bba869-bfb4-44aa-9510-6dfb82797bf5-operator-scripts\") pod \"keystone-fa91-account-create-update-k9xs4\" (UID: \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\") " pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.770476 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsqdp\" (UniqueName: \"kubernetes.io/projected/b4bba869-bfb4-44aa-9510-6dfb82797bf5-kube-api-access-zsqdp\") pod \"keystone-fa91-account-create-update-k9xs4\" (UID: \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\") " pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.771117 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc4053dd-426b-473b-9ac2-039a55ab62a7-operator-scripts\") pod \"keystone-db-create-vf6sz\" (UID: \"bc4053dd-426b-473b-9ac2-039a55ab62a7\") " pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.802591 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz9gw\" (UniqueName: \"kubernetes.io/projected/bc4053dd-426b-473b-9ac2-039a55ab62a7-kube-api-access-sz9gw\") pod \"keystone-db-create-vf6sz\" (UID: \"bc4053dd-426b-473b-9ac2-039a55ab62a7\") " pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.872584 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsqdp\" (UniqueName: \"kubernetes.io/projected/b4bba869-bfb4-44aa-9510-6dfb82797bf5-kube-api-access-zsqdp\") pod \"keystone-fa91-account-create-update-k9xs4\" (UID: \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\") " pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.872729 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4bba869-bfb4-44aa-9510-6dfb82797bf5-operator-scripts\") pod \"keystone-fa91-account-create-update-k9xs4\" (UID: \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\") " pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.873779 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4bba869-bfb4-44aa-9510-6dfb82797bf5-operator-scripts\") pod \"keystone-fa91-account-create-update-k9xs4\" (UID: \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\") " pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.892862 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsqdp\" (UniqueName: \"kubernetes.io/projected/b4bba869-bfb4-44aa-9510-6dfb82797bf5-kube-api-access-zsqdp\") pod \"keystone-fa91-account-create-update-k9xs4\" (UID: \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\") " pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:46 crc kubenswrapper[4805]: I1128 15:44:46.943101 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.032323 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.445000 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-g4grc"] Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.446631 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.450936 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-tjgr4" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.453090 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.460957 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g4grc"] Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.468930 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vf6sz"] Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.485021 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-config-data\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.485100 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-db-sync-config-data\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.485128 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pksgp\" (UniqueName: \"kubernetes.io/projected/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-kube-api-access-pksgp\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.485251 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-combined-ca-bundle\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.587417 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-combined-ca-bundle\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.587560 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-config-data\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.587594 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fa91-account-create-update-k9xs4"] Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.587621 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pksgp\" (UniqueName: \"kubernetes.io/projected/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-kube-api-access-pksgp\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.587709 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-db-sync-config-data\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.594229 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-config-data\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.594741 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-combined-ca-bundle\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.594845 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-db-sync-config-data\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: W1128 15:44:47.596323 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4bba869_bfb4_44aa_9510_6dfb82797bf5.slice/crio-b447edc0b82cf5a65bdd92ed33d7122ab899995713a082b69fb41a35a57ff026 WatchSource:0}: Error finding container b447edc0b82cf5a65bdd92ed33d7122ab899995713a082b69fb41a35a57ff026: Status 404 returned error can't find the container with id b447edc0b82cf5a65bdd92ed33d7122ab899995713a082b69fb41a35a57ff026 Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.606228 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pksgp\" (UniqueName: \"kubernetes.io/projected/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-kube-api-access-pksgp\") pod \"glance-db-sync-g4grc\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.778411 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vf6sz" event={"ID":"bc4053dd-426b-473b-9ac2-039a55ab62a7","Type":"ContainerStarted","Data":"165b9749994c0b329221bf6754fc0338f2accf4df65b8eaf405f46a8898dbdaf"} Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.778469 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vf6sz" event={"ID":"bc4053dd-426b-473b-9ac2-039a55ab62a7","Type":"ContainerStarted","Data":"2b18c6784f5bfdc0ddff9a47282d96f15779dcf29577ac01be662e9c5d91c8d8"} Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.778465 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g4grc" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.780531 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fa91-account-create-update-k9xs4" event={"ID":"b4bba869-bfb4-44aa-9510-6dfb82797bf5","Type":"ContainerStarted","Data":"4d4f71d0ee96e28e633cd614ac2ab22db848b839df7b6c625454acce90e318e8"} Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.780560 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fa91-account-create-update-k9xs4" event={"ID":"b4bba869-bfb4-44aa-9510-6dfb82797bf5","Type":"ContainerStarted","Data":"b447edc0b82cf5a65bdd92ed33d7122ab899995713a082b69fb41a35a57ff026"} Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.820807 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-fa91-account-create-update-k9xs4" podStartSLOduration=1.820785965 podStartE2EDuration="1.820785965s" podCreationTimestamp="2025-11-28 15:44:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:44:47.813001223 +0000 UTC m=+1114.862792534" watchObservedRunningTime="2025-11-28 15:44:47.820785965 +0000 UTC m=+1114.870577286" Nov 28 15:44:47 crc kubenswrapper[4805]: I1128 15:44:47.821165 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-vf6sz" podStartSLOduration=1.821156775 podStartE2EDuration="1.821156775s" podCreationTimestamp="2025-11-28 15:44:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:44:47.796314159 +0000 UTC m=+1114.846105490" watchObservedRunningTime="2025-11-28 15:44:47.821156775 +0000 UTC m=+1114.870948096" Nov 28 15:44:48 crc kubenswrapper[4805]: I1128 15:44:48.306244 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g4grc"] Nov 28 15:44:48 crc kubenswrapper[4805]: I1128 15:44:48.787403 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g4grc" event={"ID":"6b75302a-7cbf-4b5a-ad36-f6cdabe27221","Type":"ContainerStarted","Data":"9f90c98a3ba5703829d11528233e44b153fd0bd6c85509b8ae2a24b66f092aed"} Nov 28 15:44:48 crc kubenswrapper[4805]: I1128 15:44:48.789172 4805 generic.go:334] "Generic (PLEG): container finished" podID="bc4053dd-426b-473b-9ac2-039a55ab62a7" containerID="165b9749994c0b329221bf6754fc0338f2accf4df65b8eaf405f46a8898dbdaf" exitCode=0 Nov 28 15:44:48 crc kubenswrapper[4805]: I1128 15:44:48.789244 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vf6sz" event={"ID":"bc4053dd-426b-473b-9ac2-039a55ab62a7","Type":"ContainerDied","Data":"165b9749994c0b329221bf6754fc0338f2accf4df65b8eaf405f46a8898dbdaf"} Nov 28 15:44:48 crc kubenswrapper[4805]: I1128 15:44:48.791053 4805 generic.go:334] "Generic (PLEG): container finished" podID="b4bba869-bfb4-44aa-9510-6dfb82797bf5" containerID="4d4f71d0ee96e28e633cd614ac2ab22db848b839df7b6c625454acce90e318e8" exitCode=0 Nov 28 15:44:48 crc kubenswrapper[4805]: I1128 15:44:48.791119 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fa91-account-create-update-k9xs4" event={"ID":"b4bba869-bfb4-44aa-9510-6dfb82797bf5","Type":"ContainerDied","Data":"4d4f71d0ee96e28e633cd614ac2ab22db848b839df7b6c625454acce90e318e8"} Nov 28 15:44:48 crc kubenswrapper[4805]: E1128 15:44:48.802074 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb122aae_0e09_46b2_926c_037d25e79477.slice/crio-1ccc26baa6ad7cbc405abb6d95711fcbf099c77d7ab2dbe23883e5b60b452fc1.scope\": RecentStats: unable to find data in memory cache]" Nov 28 15:44:49 crc kubenswrapper[4805]: I1128 15:44:49.800785 4805 generic.go:334] "Generic (PLEG): container finished" podID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerID="bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52" exitCode=0 Nov 28 15:44:49 crc kubenswrapper[4805]: I1128 15:44:49.800872 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a","Type":"ContainerDied","Data":"bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52"} Nov 28 15:44:49 crc kubenswrapper[4805]: I1128 15:44:49.803908 4805 generic.go:334] "Generic (PLEG): container finished" podID="fb122aae-0e09-46b2-926c-037d25e79477" containerID="1ccc26baa6ad7cbc405abb6d95711fcbf099c77d7ab2dbe23883e5b60b452fc1" exitCode=0 Nov 28 15:44:49 crc kubenswrapper[4805]: I1128 15:44:49.803997 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb122aae-0e09-46b2-926c-037d25e79477","Type":"ContainerDied","Data":"1ccc26baa6ad7cbc405abb6d95711fcbf099c77d7ab2dbe23883e5b60b452fc1"} Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.185762 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.190794 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.240850 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsqdp\" (UniqueName: \"kubernetes.io/projected/b4bba869-bfb4-44aa-9510-6dfb82797bf5-kube-api-access-zsqdp\") pod \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\" (UID: \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\") " Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.240906 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc4053dd-426b-473b-9ac2-039a55ab62a7-operator-scripts\") pod \"bc4053dd-426b-473b-9ac2-039a55ab62a7\" (UID: \"bc4053dd-426b-473b-9ac2-039a55ab62a7\") " Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.240921 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4bba869-bfb4-44aa-9510-6dfb82797bf5-operator-scripts\") pod \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\" (UID: \"b4bba869-bfb4-44aa-9510-6dfb82797bf5\") " Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.240989 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz9gw\" (UniqueName: \"kubernetes.io/projected/bc4053dd-426b-473b-9ac2-039a55ab62a7-kube-api-access-sz9gw\") pod \"bc4053dd-426b-473b-9ac2-039a55ab62a7\" (UID: \"bc4053dd-426b-473b-9ac2-039a55ab62a7\") " Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.241395 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc4053dd-426b-473b-9ac2-039a55ab62a7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc4053dd-426b-473b-9ac2-039a55ab62a7" (UID: "bc4053dd-426b-473b-9ac2-039a55ab62a7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.241990 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bba869-bfb4-44aa-9510-6dfb82797bf5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4bba869-bfb4-44aa-9510-6dfb82797bf5" (UID: "b4bba869-bfb4-44aa-9510-6dfb82797bf5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.244853 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc4053dd-426b-473b-9ac2-039a55ab62a7-kube-api-access-sz9gw" (OuterVolumeSpecName: "kube-api-access-sz9gw") pod "bc4053dd-426b-473b-9ac2-039a55ab62a7" (UID: "bc4053dd-426b-473b-9ac2-039a55ab62a7"). InnerVolumeSpecName "kube-api-access-sz9gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.245486 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4bba869-bfb4-44aa-9510-6dfb82797bf5-kube-api-access-zsqdp" (OuterVolumeSpecName: "kube-api-access-zsqdp") pod "b4bba869-bfb4-44aa-9510-6dfb82797bf5" (UID: "b4bba869-bfb4-44aa-9510-6dfb82797bf5"). InnerVolumeSpecName "kube-api-access-zsqdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.342532 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsqdp\" (UniqueName: \"kubernetes.io/projected/b4bba869-bfb4-44aa-9510-6dfb82797bf5-kube-api-access-zsqdp\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.342587 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc4053dd-426b-473b-9ac2-039a55ab62a7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.342608 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4bba869-bfb4-44aa-9510-6dfb82797bf5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.342626 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz9gw\" (UniqueName: \"kubernetes.io/projected/bc4053dd-426b-473b-9ac2-039a55ab62a7-kube-api-access-sz9gw\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.814804 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fa91-account-create-update-k9xs4" event={"ID":"b4bba869-bfb4-44aa-9510-6dfb82797bf5","Type":"ContainerDied","Data":"b447edc0b82cf5a65bdd92ed33d7122ab899995713a082b69fb41a35a57ff026"} Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.815101 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b447edc0b82cf5a65bdd92ed33d7122ab899995713a082b69fb41a35a57ff026" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.815123 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fa91-account-create-update-k9xs4" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.817327 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vf6sz" event={"ID":"bc4053dd-426b-473b-9ac2-039a55ab62a7","Type":"ContainerDied","Data":"2b18c6784f5bfdc0ddff9a47282d96f15779dcf29577ac01be662e9c5d91c8d8"} Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.817391 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vf6sz" Nov 28 15:44:50 crc kubenswrapper[4805]: I1128 15:44:50.817391 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b18c6784f5bfdc0ddff9a47282d96f15779dcf29577ac01be662e9c5d91c8d8" Nov 28 15:44:52 crc kubenswrapper[4805]: I1128 15:44:52.847243 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a","Type":"ContainerStarted","Data":"cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9"} Nov 28 15:44:52 crc kubenswrapper[4805]: I1128 15:44:52.849209 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb122aae-0e09-46b2-926c-037d25e79477","Type":"ContainerStarted","Data":"8c7376fa9dd5297887afcd13bb656b98782355f020c1ad2e8dfa78a8d4ee511a"} Nov 28 15:44:53 crc kubenswrapper[4805]: I1128 15:44:53.867492 4805 generic.go:334] "Generic (PLEG): container finished" podID="300c24d4-04ee-4913-8e98-954c3f5c6f94" containerID="65d607a09a90f0f414b90b8d0bc2013bb6501ea2d7d313a6ba0262c0e54abf2d" exitCode=0 Nov 28 15:44:53 crc kubenswrapper[4805]: I1128 15:44:53.867578 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7f2hz" event={"ID":"300c24d4-04ee-4913-8e98-954c3f5c6f94","Type":"ContainerDied","Data":"65d607a09a90f0f414b90b8d0bc2013bb6501ea2d7d313a6ba0262c0e54abf2d"} Nov 28 15:44:53 crc kubenswrapper[4805]: I1128 15:44:53.868008 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 28 15:44:53 crc kubenswrapper[4805]: I1128 15:44:53.891532 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.738660032 podStartE2EDuration="1m1.891519366s" podCreationTimestamp="2025-11-28 15:43:52 +0000 UTC" firstStartedPulling="2025-11-28 15:43:54.311482785 +0000 UTC m=+1061.361274106" lastFinishedPulling="2025-11-28 15:44:15.464342139 +0000 UTC m=+1082.514133440" observedRunningTime="2025-11-28 15:44:53.887249549 +0000 UTC m=+1120.937040860" watchObservedRunningTime="2025-11-28 15:44:53.891519366 +0000 UTC m=+1120.941310677" Nov 28 15:44:53 crc kubenswrapper[4805]: I1128 15:44:53.921145 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.642948137 podStartE2EDuration="1m1.921126122s" podCreationTimestamp="2025-11-28 15:43:52 +0000 UTC" firstStartedPulling="2025-11-28 15:43:54.071942538 +0000 UTC m=+1061.121733849" lastFinishedPulling="2025-11-28 15:44:15.350120523 +0000 UTC m=+1082.399911834" observedRunningTime="2025-11-28 15:44:53.913139534 +0000 UTC m=+1120.962930875" watchObservedRunningTime="2025-11-28 15:44:53.921126122 +0000 UTC m=+1120.970917433" Nov 28 15:44:54 crc kubenswrapper[4805]: I1128 15:44:54.294251 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4q9c2" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" probeResult="failure" output=< Nov 28 15:44:54 crc kubenswrapper[4805]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 28 15:44:54 crc kubenswrapper[4805]: > Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.194501 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.337514 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-scripts\") pod \"300c24d4-04ee-4913-8e98-954c3f5c6f94\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.338647 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-combined-ca-bundle\") pod \"300c24d4-04ee-4913-8e98-954c3f5c6f94\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.338671 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-swiftconf\") pod \"300c24d4-04ee-4913-8e98-954c3f5c6f94\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.338745 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-ring-data-devices\") pod \"300c24d4-04ee-4913-8e98-954c3f5c6f94\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.338819 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgw28\" (UniqueName: \"kubernetes.io/projected/300c24d4-04ee-4913-8e98-954c3f5c6f94-kube-api-access-qgw28\") pod \"300c24d4-04ee-4913-8e98-954c3f5c6f94\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.338854 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/300c24d4-04ee-4913-8e98-954c3f5c6f94-etc-swift\") pod \"300c24d4-04ee-4913-8e98-954c3f5c6f94\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.338883 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-dispersionconf\") pod \"300c24d4-04ee-4913-8e98-954c3f5c6f94\" (UID: \"300c24d4-04ee-4913-8e98-954c3f5c6f94\") " Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.339766 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "300c24d4-04ee-4913-8e98-954c3f5c6f94" (UID: "300c24d4-04ee-4913-8e98-954c3f5c6f94"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.342213 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/300c24d4-04ee-4913-8e98-954c3f5c6f94-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "300c24d4-04ee-4913-8e98-954c3f5c6f94" (UID: "300c24d4-04ee-4913-8e98-954c3f5c6f94"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.345836 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/300c24d4-04ee-4913-8e98-954c3f5c6f94-kube-api-access-qgw28" (OuterVolumeSpecName: "kube-api-access-qgw28") pod "300c24d4-04ee-4913-8e98-954c3f5c6f94" (UID: "300c24d4-04ee-4913-8e98-954c3f5c6f94"). InnerVolumeSpecName "kube-api-access-qgw28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.351125 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "300c24d4-04ee-4913-8e98-954c3f5c6f94" (UID: "300c24d4-04ee-4913-8e98-954c3f5c6f94"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.373268 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "300c24d4-04ee-4913-8e98-954c3f5c6f94" (UID: "300c24d4-04ee-4913-8e98-954c3f5c6f94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.375529 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "300c24d4-04ee-4913-8e98-954c3f5c6f94" (UID: "300c24d4-04ee-4913-8e98-954c3f5c6f94"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.379622 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-scripts" (OuterVolumeSpecName: "scripts") pod "300c24d4-04ee-4913-8e98-954c3f5c6f94" (UID: "300c24d4-04ee-4913-8e98-954c3f5c6f94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.441084 4805 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/300c24d4-04ee-4913-8e98-954c3f5c6f94-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.441129 4805 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.441146 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.441157 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.441170 4805 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/300c24d4-04ee-4913-8e98-954c3f5c6f94-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.441180 4805 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/300c24d4-04ee-4913-8e98-954c3f5c6f94-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.441192 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgw28\" (UniqueName: \"kubernetes.io/projected/300c24d4-04ee-4913-8e98-954c3f5c6f94-kube-api-access-qgw28\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.885610 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7f2hz" event={"ID":"300c24d4-04ee-4913-8e98-954c3f5c6f94","Type":"ContainerDied","Data":"fcc8807240cd9ee3bfb821e919b131cb12251b53ed44a7e4e6da512446fff68e"} Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.885646 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcc8807240cd9ee3bfb821e919b131cb12251b53ed44a7e4e6da512446fff68e" Nov 28 15:44:55 crc kubenswrapper[4805]: I1128 15:44:55.885663 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7f2hz" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.155776 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.160972 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.290004 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4q9c2" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" probeResult="failure" output=< Nov 28 15:44:59 crc kubenswrapper[4805]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 28 15:44:59 crc kubenswrapper[4805]: > Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.397849 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4q9c2-config-v8z6m"] Nov 28 15:44:59 crc kubenswrapper[4805]: E1128 15:44:59.398256 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4bba869-bfb4-44aa-9510-6dfb82797bf5" containerName="mariadb-account-create-update" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.398282 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4bba869-bfb4-44aa-9510-6dfb82797bf5" containerName="mariadb-account-create-update" Nov 28 15:44:59 crc kubenswrapper[4805]: E1128 15:44:59.398300 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc4053dd-426b-473b-9ac2-039a55ab62a7" containerName="mariadb-database-create" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.398308 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc4053dd-426b-473b-9ac2-039a55ab62a7" containerName="mariadb-database-create" Nov 28 15:44:59 crc kubenswrapper[4805]: E1128 15:44:59.398335 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300c24d4-04ee-4913-8e98-954c3f5c6f94" containerName="swift-ring-rebalance" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.398344 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="300c24d4-04ee-4913-8e98-954c3f5c6f94" containerName="swift-ring-rebalance" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.398561 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4bba869-bfb4-44aa-9510-6dfb82797bf5" containerName="mariadb-account-create-update" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.398596 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc4053dd-426b-473b-9ac2-039a55ab62a7" containerName="mariadb-database-create" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.398612 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="300c24d4-04ee-4913-8e98-954c3f5c6f94" containerName="swift-ring-rebalance" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.399221 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.401304 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.422409 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4q9c2-config-v8z6m"] Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.527753 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-log-ovn\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.527830 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run-ovn\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.527855 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-scripts\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.527873 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7spk\" (UniqueName: \"kubernetes.io/projected/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-kube-api-access-z7spk\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.527913 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-additional-scripts\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.527958 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.629721 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run-ovn\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.629772 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-scripts\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.629791 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7spk\" (UniqueName: \"kubernetes.io/projected/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-kube-api-access-z7spk\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.629838 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-additional-scripts\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.629886 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.629946 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-log-ovn\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.630059 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-log-ovn\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.630058 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run-ovn\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.630834 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-additional-scripts\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.631180 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.632154 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-scripts\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.651270 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7spk\" (UniqueName: \"kubernetes.io/projected/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-kube-api-access-z7spk\") pod \"ovn-controller-4q9c2-config-v8z6m\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:44:59 crc kubenswrapper[4805]: I1128 15:44:59.723119 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.165305 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf"] Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.166832 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.170417 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.180265 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.184206 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf"] Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.342144 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9629789a-d860-4af6-a810-1409458b5839-secret-volume\") pod \"collect-profiles-29405745-tmhsf\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.342303 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9629789a-d860-4af6-a810-1409458b5839-config-volume\") pod \"collect-profiles-29405745-tmhsf\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.342336 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lnn2\" (UniqueName: \"kubernetes.io/projected/9629789a-d860-4af6-a810-1409458b5839-kube-api-access-6lnn2\") pod \"collect-profiles-29405745-tmhsf\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.444058 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lnn2\" (UniqueName: \"kubernetes.io/projected/9629789a-d860-4af6-a810-1409458b5839-kube-api-access-6lnn2\") pod \"collect-profiles-29405745-tmhsf\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.444494 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9629789a-d860-4af6-a810-1409458b5839-secret-volume\") pod \"collect-profiles-29405745-tmhsf\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.444653 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9629789a-d860-4af6-a810-1409458b5839-config-volume\") pod \"collect-profiles-29405745-tmhsf\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.445716 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9629789a-d860-4af6-a810-1409458b5839-config-volume\") pod \"collect-profiles-29405745-tmhsf\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.448823 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9629789a-d860-4af6-a810-1409458b5839-secret-volume\") pod \"collect-profiles-29405745-tmhsf\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.461903 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lnn2\" (UniqueName: \"kubernetes.io/projected/9629789a-d860-4af6-a810-1409458b5839-kube-api-access-6lnn2\") pod \"collect-profiles-29405745-tmhsf\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:00 crc kubenswrapper[4805]: I1128 15:45:00.491506 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:02 crc kubenswrapper[4805]: I1128 15:45:02.386175 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:45:02 crc kubenswrapper[4805]: I1128 15:45:02.405139 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"swift-storage-0\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " pod="openstack/swift-storage-0" Nov 28 15:45:02 crc kubenswrapper[4805]: I1128 15:45:02.554240 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 15:45:03 crc kubenswrapper[4805]: I1128 15:45:03.541256 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Nov 28 15:45:03 crc kubenswrapper[4805]: I1128 15:45:03.772899 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:45:03 crc kubenswrapper[4805]: I1128 15:45:03.775698 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:45:04 crc kubenswrapper[4805]: I1128 15:45:04.295774 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4q9c2" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" probeResult="failure" output=< Nov 28 15:45:04 crc kubenswrapper[4805]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 28 15:45:04 crc kubenswrapper[4805]: > Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.626261 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf"] Nov 28 15:45:05 crc kubenswrapper[4805]: W1128 15:45:05.641881 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9629789a_d860_4af6_a810_1409458b5839.slice/crio-80b2aad74cf4827df2fe26fa4c5df4fa4d00bbbd16d5e302c122803f6744a518 WatchSource:0}: Error finding container 80b2aad74cf4827df2fe26fa4c5df4fa4d00bbbd16d5e302c122803f6744a518: Status 404 returned error can't find the container with id 80b2aad74cf4827df2fe26fa4c5df4fa4d00bbbd16d5e302c122803f6744a518 Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.732095 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4q9c2-config-v8z6m"] Nov 28 15:45:05 crc kubenswrapper[4805]: W1128 15:45:05.737878 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod795dd848_e8b7_41ed_81c1_bc7135ad9f9e.slice/crio-b26a9bbbe5c465aee8bd37fb45e7b78bd8cc528e127017ec489ae902b64a5883 WatchSource:0}: Error finding container b26a9bbbe5c465aee8bd37fb45e7b78bd8cc528e127017ec489ae902b64a5883: Status 404 returned error can't find the container with id b26a9bbbe5c465aee8bd37fb45e7b78bd8cc528e127017ec489ae902b64a5883 Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.815951 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.967666 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g4grc" event={"ID":"6b75302a-7cbf-4b5a-ad36-f6cdabe27221","Type":"ContainerStarted","Data":"01b2f5f82ace63d29f30997e3ed6d0a996047e0556752afaa1bc33b25beb49ca"} Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.970443 4805 generic.go:334] "Generic (PLEG): container finished" podID="9629789a-d860-4af6-a810-1409458b5839" containerID="6d27a48ff0abf4b9dedb8f69db9a3f5aa29d2c0f854534386aa9397fef8b574d" exitCode=0 Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.970498 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" event={"ID":"9629789a-d860-4af6-a810-1409458b5839","Type":"ContainerDied","Data":"6d27a48ff0abf4b9dedb8f69db9a3f5aa29d2c0f854534386aa9397fef8b574d"} Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.970521 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" event={"ID":"9629789a-d860-4af6-a810-1409458b5839","Type":"ContainerStarted","Data":"80b2aad74cf4827df2fe26fa4c5df4fa4d00bbbd16d5e302c122803f6744a518"} Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.979306 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2-config-v8z6m" event={"ID":"795dd848-e8b7-41ed-81c1-bc7135ad9f9e","Type":"ContainerStarted","Data":"b26a9bbbe5c465aee8bd37fb45e7b78bd8cc528e127017ec489ae902b64a5883"} Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.980704 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"69fff0ee7eff65d669d3ec50922824c17777dd709a0a3ea5ac2e57087bb8a133"} Nov 28 15:45:05 crc kubenswrapper[4805]: I1128 15:45:05.995229 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-g4grc" podStartSLOduration=2.067479122 podStartE2EDuration="18.995212779s" podCreationTimestamp="2025-11-28 15:44:47 +0000 UTC" firstStartedPulling="2025-11-28 15:44:48.309853194 +0000 UTC m=+1115.359644505" lastFinishedPulling="2025-11-28 15:45:05.237586851 +0000 UTC m=+1132.287378162" observedRunningTime="2025-11-28 15:45:05.992443914 +0000 UTC m=+1133.042235235" watchObservedRunningTime="2025-11-28 15:45:05.995212779 +0000 UTC m=+1133.045004090" Nov 28 15:45:06 crc kubenswrapper[4805]: I1128 15:45:06.991475 4805 generic.go:334] "Generic (PLEG): container finished" podID="795dd848-e8b7-41ed-81c1-bc7135ad9f9e" containerID="5dec4cf0f7bf4573bbaa714bf92c31c74fecb1cf0fab84de687c65be40e99c1f" exitCode=0 Nov 28 15:45:06 crc kubenswrapper[4805]: I1128 15:45:06.991608 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2-config-v8z6m" event={"ID":"795dd848-e8b7-41ed-81c1-bc7135ad9f9e","Type":"ContainerDied","Data":"5dec4cf0f7bf4573bbaa714bf92c31c74fecb1cf0fab84de687c65be40e99c1f"} Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.373297 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.467874 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lnn2\" (UniqueName: \"kubernetes.io/projected/9629789a-d860-4af6-a810-1409458b5839-kube-api-access-6lnn2\") pod \"9629789a-d860-4af6-a810-1409458b5839\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.467976 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9629789a-d860-4af6-a810-1409458b5839-config-volume\") pod \"9629789a-d860-4af6-a810-1409458b5839\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.468025 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9629789a-d860-4af6-a810-1409458b5839-secret-volume\") pod \"9629789a-d860-4af6-a810-1409458b5839\" (UID: \"9629789a-d860-4af6-a810-1409458b5839\") " Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.468621 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9629789a-d860-4af6-a810-1409458b5839-config-volume" (OuterVolumeSpecName: "config-volume") pod "9629789a-d860-4af6-a810-1409458b5839" (UID: "9629789a-d860-4af6-a810-1409458b5839"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.477902 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9629789a-d860-4af6-a810-1409458b5839-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9629789a-d860-4af6-a810-1409458b5839" (UID: "9629789a-d860-4af6-a810-1409458b5839"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.479550 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9629789a-d860-4af6-a810-1409458b5839-kube-api-access-6lnn2" (OuterVolumeSpecName: "kube-api-access-6lnn2") pod "9629789a-d860-4af6-a810-1409458b5839" (UID: "9629789a-d860-4af6-a810-1409458b5839"). InnerVolumeSpecName "kube-api-access-6lnn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.570290 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9629789a-d860-4af6-a810-1409458b5839-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.570461 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9629789a-d860-4af6-a810-1409458b5839-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:07 crc kubenswrapper[4805]: I1128 15:45:07.570524 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lnn2\" (UniqueName: \"kubernetes.io/projected/9629789a-d860-4af6-a810-1409458b5839-kube-api-access-6lnn2\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.002103 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" event={"ID":"9629789a-d860-4af6-a810-1409458b5839","Type":"ContainerDied","Data":"80b2aad74cf4827df2fe26fa4c5df4fa4d00bbbd16d5e302c122803f6744a518"} Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.002776 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80b2aad74cf4827df2fe26fa4c5df4fa4d00bbbd16d5e302c122803f6744a518" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.002133 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-tmhsf" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.008793 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"663742bde9628596b637c938ce19b2d482d18491b8629d57987967e1b1b24891"} Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.008965 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"d259dddfea43ce8231ff4ea937ce34f8e372f66bcd4c5effa7aec7ce0ffa8390"} Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.009028 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"3ead42c6ea605c5d27888381393dc51bf09a9814de73cd305dbca46329a8a17d"} Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.009093 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"cbabf490d47a58ab2c8c9f6eacb4cf048920fd4314a573b001947462a0de16b4"} Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.283374 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.382663 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-additional-scripts\") pod \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.382861 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-log-ovn\") pod \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.382884 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run-ovn\") pod \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.382919 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-scripts\") pod \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.382929 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "795dd848-e8b7-41ed-81c1-bc7135ad9f9e" (UID: "795dd848-e8b7-41ed-81c1-bc7135ad9f9e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.382982 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run\") pod \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.383040 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7spk\" (UniqueName: \"kubernetes.io/projected/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-kube-api-access-z7spk\") pod \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\" (UID: \"795dd848-e8b7-41ed-81c1-bc7135ad9f9e\") " Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.382976 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "795dd848-e8b7-41ed-81c1-bc7135ad9f9e" (UID: "795dd848-e8b7-41ed-81c1-bc7135ad9f9e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.383049 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run" (OuterVolumeSpecName: "var-run") pod "795dd848-e8b7-41ed-81c1-bc7135ad9f9e" (UID: "795dd848-e8b7-41ed-81c1-bc7135ad9f9e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.383349 4805 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.383388 4805 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.383399 4805 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.383566 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "795dd848-e8b7-41ed-81c1-bc7135ad9f9e" (UID: "795dd848-e8b7-41ed-81c1-bc7135ad9f9e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.383805 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-scripts" (OuterVolumeSpecName: "scripts") pod "795dd848-e8b7-41ed-81c1-bc7135ad9f9e" (UID: "795dd848-e8b7-41ed-81c1-bc7135ad9f9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.387334 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-kube-api-access-z7spk" (OuterVolumeSpecName: "kube-api-access-z7spk") pod "795dd848-e8b7-41ed-81c1-bc7135ad9f9e" (UID: "795dd848-e8b7-41ed-81c1-bc7135ad9f9e"). InnerVolumeSpecName "kube-api-access-z7spk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.485034 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7spk\" (UniqueName: \"kubernetes.io/projected/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-kube-api-access-z7spk\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.485070 4805 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:08 crc kubenswrapper[4805]: I1128 15:45:08.485081 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/795dd848-e8b7-41ed-81c1-bc7135ad9f9e-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.021509 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2-config-v8z6m" event={"ID":"795dd848-e8b7-41ed-81c1-bc7135ad9f9e","Type":"ContainerDied","Data":"b26a9bbbe5c465aee8bd37fb45e7b78bd8cc528e127017ec489ae902b64a5883"} Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.021836 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b26a9bbbe5c465aee8bd37fb45e7b78bd8cc528e127017ec489ae902b64a5883" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.021569 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2-config-v8z6m" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.303287 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-4q9c2" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.389894 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4q9c2-config-v8z6m"] Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.397053 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4q9c2-config-v8z6m"] Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.517470 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4q9c2-config-6spbh"] Nov 28 15:45:09 crc kubenswrapper[4805]: E1128 15:45:09.519835 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795dd848-e8b7-41ed-81c1-bc7135ad9f9e" containerName="ovn-config" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.519873 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="795dd848-e8b7-41ed-81c1-bc7135ad9f9e" containerName="ovn-config" Nov 28 15:45:09 crc kubenswrapper[4805]: E1128 15:45:09.519891 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9629789a-d860-4af6-a810-1409458b5839" containerName="collect-profiles" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.519901 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9629789a-d860-4af6-a810-1409458b5839" containerName="collect-profiles" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.520181 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9629789a-d860-4af6-a810-1409458b5839" containerName="collect-profiles" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.520211 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="795dd848-e8b7-41ed-81c1-bc7135ad9f9e" containerName="ovn-config" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.520914 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.524136 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.524582 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4q9c2-config-6spbh"] Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.602164 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run-ovn\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.602216 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.602270 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-scripts\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.602299 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-additional-scripts\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.602415 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-log-ovn\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.602451 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sgz8\" (UniqueName: \"kubernetes.io/projected/ee9740d4-8948-49c7-a093-663b0c6c438f-kube-api-access-4sgz8\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.704055 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-log-ovn\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.704107 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sgz8\" (UniqueName: \"kubernetes.io/projected/ee9740d4-8948-49c7-a093-663b0c6c438f-kube-api-access-4sgz8\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.704139 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run-ovn\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.704178 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.704213 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-scripts\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.704241 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-additional-scripts\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.705369 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-additional-scripts\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.705610 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run-ovn\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.705663 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-log-ovn\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.705702 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.707857 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-scripts\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.729075 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sgz8\" (UniqueName: \"kubernetes.io/projected/ee9740d4-8948-49c7-a093-663b0c6c438f-kube-api-access-4sgz8\") pod \"ovn-controller-4q9c2-config-6spbh\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:09 crc kubenswrapper[4805]: I1128 15:45:09.859139 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:10 crc kubenswrapper[4805]: I1128 15:45:10.037036 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"7c0cf000e71a5c9d2c9ce21d71ecb228ec004603af520ea9f4b7232773601e18"} Nov 28 15:45:10 crc kubenswrapper[4805]: I1128 15:45:10.037088 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"10fe5f001457130421ee90b15666db980a7890c30af98e597a996d50a50c4e9a"} Nov 28 15:45:10 crc kubenswrapper[4805]: I1128 15:45:10.037100 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"831f948c09940fb083fcf963fed4d7066d04ebc8f2adc914bc711740e0fcdf49"} Nov 28 15:45:10 crc kubenswrapper[4805]: I1128 15:45:10.389380 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4q9c2-config-6spbh"] Nov 28 15:45:10 crc kubenswrapper[4805]: W1128 15:45:10.392391 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee9740d4_8948_49c7_a093_663b0c6c438f.slice/crio-5a3e1c3a3613e9642a2e6f54efb9601720f71bd0492bb8bfcfd3d10cf274efbd WatchSource:0}: Error finding container 5a3e1c3a3613e9642a2e6f54efb9601720f71bd0492bb8bfcfd3d10cf274efbd: Status 404 returned error can't find the container with id 5a3e1c3a3613e9642a2e6f54efb9601720f71bd0492bb8bfcfd3d10cf274efbd Nov 28 15:45:11 crc kubenswrapper[4805]: I1128 15:45:11.046186 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2-config-6spbh" event={"ID":"ee9740d4-8948-49c7-a093-663b0c6c438f","Type":"ContainerStarted","Data":"5a3e1c3a3613e9642a2e6f54efb9601720f71bd0492bb8bfcfd3d10cf274efbd"} Nov 28 15:45:11 crc kubenswrapper[4805]: I1128 15:45:11.049893 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"fbaa7d95c8f6e7a6d621aa209aea65439fd1c81e062456fd47f282542a39aea3"} Nov 28 15:45:11 crc kubenswrapper[4805]: I1128 15:45:11.060390 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:45:11 crc kubenswrapper[4805]: I1128 15:45:11.060451 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:45:11 crc kubenswrapper[4805]: I1128 15:45:11.060492 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:45:11 crc kubenswrapper[4805]: I1128 15:45:11.061098 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ada3fc8f488f7799cae44933973db15a3177119e0d7d2d4556df785008a6700d"} pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:45:11 crc kubenswrapper[4805]: I1128 15:45:11.061158 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" containerID="cri-o://ada3fc8f488f7799cae44933973db15a3177119e0d7d2d4556df785008a6700d" gracePeriod=600 Nov 28 15:45:11 crc kubenswrapper[4805]: I1128 15:45:11.216811 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="795dd848-e8b7-41ed-81c1-bc7135ad9f9e" path="/var/lib/kubelet/pods/795dd848-e8b7-41ed-81c1-bc7135ad9f9e/volumes" Nov 28 15:45:12 crc kubenswrapper[4805]: I1128 15:45:12.061732 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"a92aa0f7330344e2f61dbf42e3f768648b41c7554cad6173713ec787c88993fc"} Nov 28 15:45:12 crc kubenswrapper[4805]: I1128 15:45:12.064297 4805 generic.go:334] "Generic (PLEG): container finished" podID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerID="ada3fc8f488f7799cae44933973db15a3177119e0d7d2d4556df785008a6700d" exitCode=0 Nov 28 15:45:12 crc kubenswrapper[4805]: I1128 15:45:12.064399 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerDied","Data":"ada3fc8f488f7799cae44933973db15a3177119e0d7d2d4556df785008a6700d"} Nov 28 15:45:12 crc kubenswrapper[4805]: I1128 15:45:12.064460 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"f2d2222cce3bf61f8537971c6c7eb69d6ee1501220244e7d4ba5697a12f1d2d5"} Nov 28 15:45:12 crc kubenswrapper[4805]: I1128 15:45:12.064478 4805 scope.go:117] "RemoveContainer" containerID="9401a6879e1fcc43d0b27657309da77bf635d82861975cc53606a87548ec258e" Nov 28 15:45:12 crc kubenswrapper[4805]: I1128 15:45:12.069688 4805 generic.go:334] "Generic (PLEG): container finished" podID="ee9740d4-8948-49c7-a093-663b0c6c438f" containerID="14a51093a1b996f3d81f038eac81b07209514a1e4e613ae566cfbba1064f591f" exitCode=0 Nov 28 15:45:12 crc kubenswrapper[4805]: I1128 15:45:12.069724 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2-config-6spbh" event={"ID":"ee9740d4-8948-49c7-a093-663b0c6c438f","Type":"ContainerDied","Data":"14a51093a1b996f3d81f038eac81b07209514a1e4e613ae566cfbba1064f591f"} Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.087131 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"68a336481b5b62ed5a8243f246583d3e81d48c711638463168818d7ae197b2fe"} Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.087929 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"4511ff31bb1feea8dab2de6f562fa4e27fc7ba92977a8fbc6bf26f4714ba0be0"} Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.087953 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"7968f4c1d7d9d723379b1a1881a1e60cc6af6ce686c094d24597209d9d9c45cd"} Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.374630 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.472317 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run-ovn\") pod \"ee9740d4-8948-49c7-a093-663b0c6c438f\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.472719 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run\") pod \"ee9740d4-8948-49c7-a093-663b0c6c438f\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.472779 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-scripts\") pod \"ee9740d4-8948-49c7-a093-663b0c6c438f\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.472881 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-additional-scripts\") pod \"ee9740d4-8948-49c7-a093-663b0c6c438f\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.472908 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sgz8\" (UniqueName: \"kubernetes.io/projected/ee9740d4-8948-49c7-a093-663b0c6c438f-kube-api-access-4sgz8\") pod \"ee9740d4-8948-49c7-a093-663b0c6c438f\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.472933 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-log-ovn\") pod \"ee9740d4-8948-49c7-a093-663b0c6c438f\" (UID: \"ee9740d4-8948-49c7-a093-663b0c6c438f\") " Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.472468 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "ee9740d4-8948-49c7-a093-663b0c6c438f" (UID: "ee9740d4-8948-49c7-a093-663b0c6c438f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.473316 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "ee9740d4-8948-49c7-a093-663b0c6c438f" (UID: "ee9740d4-8948-49c7-a093-663b0c6c438f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.473377 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run" (OuterVolumeSpecName: "var-run") pod "ee9740d4-8948-49c7-a093-663b0c6c438f" (UID: "ee9740d4-8948-49c7-a093-663b0c6c438f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.473890 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "ee9740d4-8948-49c7-a093-663b0c6c438f" (UID: "ee9740d4-8948-49c7-a093-663b0c6c438f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.474424 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-scripts" (OuterVolumeSpecName: "scripts") pod "ee9740d4-8948-49c7-a093-663b0c6c438f" (UID: "ee9740d4-8948-49c7-a093-663b0c6c438f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.478543 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9740d4-8948-49c7-a093-663b0c6c438f-kube-api-access-4sgz8" (OuterVolumeSpecName: "kube-api-access-4sgz8") pod "ee9740d4-8948-49c7-a093-663b0c6c438f" (UID: "ee9740d4-8948-49c7-a093-663b0c6c438f"). InnerVolumeSpecName "kube-api-access-4sgz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.542335 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.575308 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sgz8\" (UniqueName: \"kubernetes.io/projected/ee9740d4-8948-49c7-a093-663b0c6c438f-kube-api-access-4sgz8\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.575344 4805 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.575369 4805 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.575381 4805 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee9740d4-8948-49c7-a093-663b0c6c438f-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.575392 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.575403 4805 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9740d4-8948-49c7-a093-663b0c6c438f-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.938092 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-jfqsr"] Nov 28 15:45:13 crc kubenswrapper[4805]: E1128 15:45:13.938512 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9740d4-8948-49c7-a093-663b0c6c438f" containerName="ovn-config" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.938529 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9740d4-8948-49c7-a093-663b0c6c438f" containerName="ovn-config" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.938733 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9740d4-8948-49c7-a093-663b0c6c438f" containerName="ovn-config" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.939343 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.946501 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-72ad-account-create-update-66ckj"] Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.947725 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.951945 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.968546 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-72ad-account-create-update-66ckj"] Nov 28 15:45:13 crc kubenswrapper[4805]: I1128 15:45:13.996819 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jfqsr"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.005918 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s72b9\" (UniqueName: \"kubernetes.io/projected/19093324-6f6b-47d3-a08d-051d59136892-kube-api-access-s72b9\") pod \"cinder-db-create-jfqsr\" (UID: \"19093324-6f6b-47d3-a08d-051d59136892\") " pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.006036 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19093324-6f6b-47d3-a08d-051d59136892-operator-scripts\") pod \"cinder-db-create-jfqsr\" (UID: \"19093324-6f6b-47d3-a08d-051d59136892\") " pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.103350 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2-config-6spbh" event={"ID":"ee9740d4-8948-49c7-a093-663b0c6c438f","Type":"ContainerDied","Data":"5a3e1c3a3613e9642a2e6f54efb9601720f71bd0492bb8bfcfd3d10cf274efbd"} Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.103412 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a3e1c3a3613e9642a2e6f54efb9601720f71bd0492bb8bfcfd3d10cf274efbd" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.103504 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2-config-6spbh" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.108328 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s72b9\" (UniqueName: \"kubernetes.io/projected/19093324-6f6b-47d3-a08d-051d59136892-kube-api-access-s72b9\") pod \"cinder-db-create-jfqsr\" (UID: \"19093324-6f6b-47d3-a08d-051d59136892\") " pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.108495 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxqnw\" (UniqueName: \"kubernetes.io/projected/09c6b57a-16a1-465d-b883-aa6bea6756c5-kube-api-access-xxqnw\") pod \"cinder-72ad-account-create-update-66ckj\" (UID: \"09c6b57a-16a1-465d-b883-aa6bea6756c5\") " pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.108542 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19093324-6f6b-47d3-a08d-051d59136892-operator-scripts\") pod \"cinder-db-create-jfqsr\" (UID: \"19093324-6f6b-47d3-a08d-051d59136892\") " pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.108603 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09c6b57a-16a1-465d-b883-aa6bea6756c5-operator-scripts\") pod \"cinder-72ad-account-create-update-66ckj\" (UID: \"09c6b57a-16a1-465d-b883-aa6bea6756c5\") " pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.109684 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19093324-6f6b-47d3-a08d-051d59136892-operator-scripts\") pod \"cinder-db-create-jfqsr\" (UID: \"19093324-6f6b-47d3-a08d-051d59136892\") " pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.111887 4805 generic.go:334] "Generic (PLEG): container finished" podID="6b75302a-7cbf-4b5a-ad36-f6cdabe27221" containerID="01b2f5f82ace63d29f30997e3ed6d0a996047e0556752afaa1bc33b25beb49ca" exitCode=0 Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.111943 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g4grc" event={"ID":"6b75302a-7cbf-4b5a-ad36-f6cdabe27221","Type":"ContainerDied","Data":"01b2f5f82ace63d29f30997e3ed6d0a996047e0556752afaa1bc33b25beb49ca"} Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.132882 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-dnbhb"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.134528 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.148039 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s72b9\" (UniqueName: \"kubernetes.io/projected/19093324-6f6b-47d3-a08d-051d59136892-kube-api-access-s72b9\") pod \"cinder-db-create-jfqsr\" (UID: \"19093324-6f6b-47d3-a08d-051d59136892\") " pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.168273 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dnbhb"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.210457 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c308b8e2-19af-4cfc-9a06-487497e95198-operator-scripts\") pod \"barbican-db-create-dnbhb\" (UID: \"c308b8e2-19af-4cfc-9a06-487497e95198\") " pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.210494 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjd4f\" (UniqueName: \"kubernetes.io/projected/c308b8e2-19af-4cfc-9a06-487497e95198-kube-api-access-vjd4f\") pod \"barbican-db-create-dnbhb\" (UID: \"c308b8e2-19af-4cfc-9a06-487497e95198\") " pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.210531 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09c6b57a-16a1-465d-b883-aa6bea6756c5-operator-scripts\") pod \"cinder-72ad-account-create-update-66ckj\" (UID: \"09c6b57a-16a1-465d-b883-aa6bea6756c5\") " pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.210619 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxqnw\" (UniqueName: \"kubernetes.io/projected/09c6b57a-16a1-465d-b883-aa6bea6756c5-kube-api-access-xxqnw\") pod \"cinder-72ad-account-create-update-66ckj\" (UID: \"09c6b57a-16a1-465d-b883-aa6bea6756c5\") " pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.211480 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09c6b57a-16a1-465d-b883-aa6bea6756c5-operator-scripts\") pod \"cinder-72ad-account-create-update-66ckj\" (UID: \"09c6b57a-16a1-465d-b883-aa6bea6756c5\") " pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.213979 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8232-account-create-update-tg9lf"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.215295 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.218686 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.254022 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxqnw\" (UniqueName: \"kubernetes.io/projected/09c6b57a-16a1-465d-b883-aa6bea6756c5-kube-api-access-xxqnw\") pod \"cinder-72ad-account-create-update-66ckj\" (UID: \"09c6b57a-16a1-465d-b883-aa6bea6756c5\") " pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.259413 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.266992 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.277879 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8232-account-create-update-tg9lf"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.293066 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ct546"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.294154 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.299943 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fzt55" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.299979 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.300687 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.300836 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.313090 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c308b8e2-19af-4cfc-9a06-487497e95198-operator-scripts\") pod \"barbican-db-create-dnbhb\" (UID: \"c308b8e2-19af-4cfc-9a06-487497e95198\") " pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.313145 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjd4f\" (UniqueName: \"kubernetes.io/projected/c308b8e2-19af-4cfc-9a06-487497e95198-kube-api-access-vjd4f\") pod \"barbican-db-create-dnbhb\" (UID: \"c308b8e2-19af-4cfc-9a06-487497e95198\") " pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.313180 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgkd9\" (UniqueName: \"kubernetes.io/projected/3c642d47-f1ea-4254-a76e-5ebe959da352-kube-api-access-vgkd9\") pod \"barbican-8232-account-create-update-tg9lf\" (UID: \"3c642d47-f1ea-4254-a76e-5ebe959da352\") " pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.313344 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c642d47-f1ea-4254-a76e-5ebe959da352-operator-scripts\") pod \"barbican-8232-account-create-update-tg9lf\" (UID: \"3c642d47-f1ea-4254-a76e-5ebe959da352\") " pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.315031 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c308b8e2-19af-4cfc-9a06-487497e95198-operator-scripts\") pod \"barbican-db-create-dnbhb\" (UID: \"c308b8e2-19af-4cfc-9a06-487497e95198\") " pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.332536 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ct546"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.335463 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjd4f\" (UniqueName: \"kubernetes.io/projected/c308b8e2-19af-4cfc-9a06-487497e95198-kube-api-access-vjd4f\") pod \"barbican-db-create-dnbhb\" (UID: \"c308b8e2-19af-4cfc-9a06-487497e95198\") " pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.373170 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-1f9e-account-create-update-c5tc4"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.383626 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.386802 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.389976 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-1f9e-account-create-update-c5tc4"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.414801 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgkd9\" (UniqueName: \"kubernetes.io/projected/3c642d47-f1ea-4254-a76e-5ebe959da352-kube-api-access-vgkd9\") pod \"barbican-8232-account-create-update-tg9lf\" (UID: \"3c642d47-f1ea-4254-a76e-5ebe959da352\") " pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.414871 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfw26\" (UniqueName: \"kubernetes.io/projected/7ac018eb-cb26-4d12-ac87-c200b20a9327-kube-api-access-gfw26\") pod \"keystone-db-sync-ct546\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.414908 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-config-data\") pod \"keystone-db-sync-ct546\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.414980 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-combined-ca-bundle\") pod \"keystone-db-sync-ct546\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.415020 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c642d47-f1ea-4254-a76e-5ebe959da352-operator-scripts\") pod \"barbican-8232-account-create-update-tg9lf\" (UID: \"3c642d47-f1ea-4254-a76e-5ebe959da352\") " pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.415847 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c642d47-f1ea-4254-a76e-5ebe959da352-operator-scripts\") pod \"barbican-8232-account-create-update-tg9lf\" (UID: \"3c642d47-f1ea-4254-a76e-5ebe959da352\") " pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.449730 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgkd9\" (UniqueName: \"kubernetes.io/projected/3c642d47-f1ea-4254-a76e-5ebe959da352-kube-api-access-vgkd9\") pod \"barbican-8232-account-create-update-tg9lf\" (UID: \"3c642d47-f1ea-4254-a76e-5ebe959da352\") " pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.457405 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-bzx4d"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.458558 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.476155 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bzx4d"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.484271 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4q9c2-config-6spbh"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.498740 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4q9c2-config-6spbh"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.508346 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.517018 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlbfr\" (UniqueName: \"kubernetes.io/projected/7a29bf56-ccc6-4b62-a4b7-da0093d80478-kube-api-access-zlbfr\") pod \"neutron-1f9e-account-create-update-c5tc4\" (UID: \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\") " pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.517079 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52942\" (UniqueName: \"kubernetes.io/projected/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-kube-api-access-52942\") pod \"neutron-db-create-bzx4d\" (UID: \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\") " pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.517165 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfw26\" (UniqueName: \"kubernetes.io/projected/7ac018eb-cb26-4d12-ac87-c200b20a9327-kube-api-access-gfw26\") pod \"keystone-db-sync-ct546\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.517203 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-config-data\") pod \"keystone-db-sync-ct546\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.517230 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a29bf56-ccc6-4b62-a4b7-da0093d80478-operator-scripts\") pod \"neutron-1f9e-account-create-update-c5tc4\" (UID: \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\") " pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.517298 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-combined-ca-bundle\") pod \"keystone-db-sync-ct546\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.517341 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-operator-scripts\") pod \"neutron-db-create-bzx4d\" (UID: \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\") " pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.522045 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-config-data\") pod \"keystone-db-sync-ct546\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.533252 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.535300 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-combined-ca-bundle\") pod \"keystone-db-sync-ct546\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.541502 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfw26\" (UniqueName: \"kubernetes.io/projected/7ac018eb-cb26-4d12-ac87-c200b20a9327-kube-api-access-gfw26\") pod \"keystone-db-sync-ct546\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.621653 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a29bf56-ccc6-4b62-a4b7-da0093d80478-operator-scripts\") pod \"neutron-1f9e-account-create-update-c5tc4\" (UID: \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\") " pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.621732 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-operator-scripts\") pod \"neutron-db-create-bzx4d\" (UID: \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\") " pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.621760 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlbfr\" (UniqueName: \"kubernetes.io/projected/7a29bf56-ccc6-4b62-a4b7-da0093d80478-kube-api-access-zlbfr\") pod \"neutron-1f9e-account-create-update-c5tc4\" (UID: \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\") " pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.621787 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52942\" (UniqueName: \"kubernetes.io/projected/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-kube-api-access-52942\") pod \"neutron-db-create-bzx4d\" (UID: \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\") " pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.622461 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a29bf56-ccc6-4b62-a4b7-da0093d80478-operator-scripts\") pod \"neutron-1f9e-account-create-update-c5tc4\" (UID: \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\") " pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.622666 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-operator-scripts\") pod \"neutron-db-create-bzx4d\" (UID: \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\") " pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.642939 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52942\" (UniqueName: \"kubernetes.io/projected/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-kube-api-access-52942\") pod \"neutron-db-create-bzx4d\" (UID: \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\") " pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.644269 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlbfr\" (UniqueName: \"kubernetes.io/projected/7a29bf56-ccc6-4b62-a4b7-da0093d80478-kube-api-access-zlbfr\") pod \"neutron-1f9e-account-create-update-c5tc4\" (UID: \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\") " pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.748059 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.754195 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.759702 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-72ad-account-create-update-66ckj"] Nov 28 15:45:14 crc kubenswrapper[4805]: I1128 15:45:14.779646 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.030697 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jfqsr"] Nov 28 15:45:15 crc kubenswrapper[4805]: W1128 15:45:15.030871 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19093324_6f6b_47d3_a08d_051d59136892.slice/crio-9d047158c51e3bcb3f4c4a4091f7c4369f41e0387063e21a87dcfcc1102b6704 WatchSource:0}: Error finding container 9d047158c51e3bcb3f4c4a4091f7c4369f41e0387063e21a87dcfcc1102b6704: Status 404 returned error can't find the container with id 9d047158c51e3bcb3f4c4a4091f7c4369f41e0387063e21a87dcfcc1102b6704 Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.107453 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dnbhb"] Nov 28 15:45:15 crc kubenswrapper[4805]: W1128 15:45:15.117729 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc308b8e2_19af_4cfc_9a06_487497e95198.slice/crio-a274179c43990cd11524eacb2c0b21565d486ccc99011fd110aad43d08e0f146 WatchSource:0}: Error finding container a274179c43990cd11524eacb2c0b21565d486ccc99011fd110aad43d08e0f146: Status 404 returned error can't find the container with id a274179c43990cd11524eacb2c0b21565d486ccc99011fd110aad43d08e0f146 Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.119924 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ct546"] Nov 28 15:45:15 crc kubenswrapper[4805]: W1128 15:45:15.123510 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ac018eb_cb26_4d12_ac87_c200b20a9327.slice/crio-3c08b756ec5332e4fcb92d5c64d172647c637bd580632d09a5aa967347cef0f8 WatchSource:0}: Error finding container 3c08b756ec5332e4fcb92d5c64d172647c637bd580632d09a5aa967347cef0f8: Status 404 returned error can't find the container with id 3c08b756ec5332e4fcb92d5c64d172647c637bd580632d09a5aa967347cef0f8 Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.150889 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-72ad-account-create-update-66ckj" event={"ID":"09c6b57a-16a1-465d-b883-aa6bea6756c5","Type":"ContainerStarted","Data":"34e1bd466c219c7477ab72652d75dbcb00f7e2a2578eb875c677493abc6d67a4"} Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.153374 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jfqsr" event={"ID":"19093324-6f6b-47d3-a08d-051d59136892","Type":"ContainerStarted","Data":"9d047158c51e3bcb3f4c4a4091f7c4369f41e0387063e21a87dcfcc1102b6704"} Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.154730 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ct546" event={"ID":"7ac018eb-cb26-4d12-ac87-c200b20a9327","Type":"ContainerStarted","Data":"3c08b756ec5332e4fcb92d5c64d172647c637bd580632d09a5aa967347cef0f8"} Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.157688 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dnbhb" event={"ID":"c308b8e2-19af-4cfc-9a06-487497e95198","Type":"ContainerStarted","Data":"a274179c43990cd11524eacb2c0b21565d486ccc99011fd110aad43d08e0f146"} Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.191846 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8232-account-create-update-tg9lf"] Nov 28 15:45:15 crc kubenswrapper[4805]: W1128 15:45:15.193759 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c642d47_f1ea_4254_a76e_5ebe959da352.slice/crio-0e6c01cb376d01229e815d582c24d17503b80982036fc45630f85e20467f0035 WatchSource:0}: Error finding container 0e6c01cb376d01229e815d582c24d17503b80982036fc45630f85e20467f0035: Status 404 returned error can't find the container with id 0e6c01cb376d01229e815d582c24d17503b80982036fc45630f85e20467f0035 Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.245762 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9740d4-8948-49c7-a093-663b0c6c438f" path="/var/lib/kubelet/pods/ee9740d4-8948-49c7-a093-663b0c6c438f/volumes" Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.400045 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bzx4d"] Nov 28 15:45:15 crc kubenswrapper[4805]: I1128 15:45:15.422374 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-1f9e-account-create-update-c5tc4"] Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.024491 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g4grc" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.167012 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g4grc" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.167002 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g4grc" event={"ID":"6b75302a-7cbf-4b5a-ad36-f6cdabe27221","Type":"ContainerDied","Data":"9f90c98a3ba5703829d11528233e44b153fd0bd6c85509b8ae2a24b66f092aed"} Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.167162 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f90c98a3ba5703829d11528233e44b153fd0bd6c85509b8ae2a24b66f092aed" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.174227 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"7bc27ced72f3c03367ba7ebd0e3bbe80199cea1ba0a0d9701c61d76ee2259b8b"} Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.177317 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bzx4d" event={"ID":"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce","Type":"ContainerStarted","Data":"7f5de6ea1047902dd8235ade08906548f8ace0bc4a666ce5c9c30355d4085975"} Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.180602 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8232-account-create-update-tg9lf" event={"ID":"3c642d47-f1ea-4254-a76e-5ebe959da352","Type":"ContainerStarted","Data":"0e6c01cb376d01229e815d582c24d17503b80982036fc45630f85e20467f0035"} Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.180826 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-config-data\") pod \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.180869 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pksgp\" (UniqueName: \"kubernetes.io/projected/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-kube-api-access-pksgp\") pod \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.180929 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-db-sync-config-data\") pod \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.181035 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-combined-ca-bundle\") pod \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\" (UID: \"6b75302a-7cbf-4b5a-ad36-f6cdabe27221\") " Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.184282 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-1f9e-account-create-update-c5tc4" event={"ID":"7a29bf56-ccc6-4b62-a4b7-da0093d80478","Type":"ContainerStarted","Data":"a9c4809489cbcdf21a54b37bb58a4528f6c3e7fb82b8b489d0d2c53019a4d184"} Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.190793 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6b75302a-7cbf-4b5a-ad36-f6cdabe27221" (UID: "6b75302a-7cbf-4b5a-ad36-f6cdabe27221"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.192238 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-kube-api-access-pksgp" (OuterVolumeSpecName: "kube-api-access-pksgp") pod "6b75302a-7cbf-4b5a-ad36-f6cdabe27221" (UID: "6b75302a-7cbf-4b5a-ad36-f6cdabe27221"). InnerVolumeSpecName "kube-api-access-pksgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.230531 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b75302a-7cbf-4b5a-ad36-f6cdabe27221" (UID: "6b75302a-7cbf-4b5a-ad36-f6cdabe27221"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.262520 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-config-data" (OuterVolumeSpecName: "config-data") pod "6b75302a-7cbf-4b5a-ad36-f6cdabe27221" (UID: "6b75302a-7cbf-4b5a-ad36-f6cdabe27221"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.283557 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.283889 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.283950 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pksgp\" (UniqueName: \"kubernetes.io/projected/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-kube-api-access-pksgp\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.284013 4805 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b75302a-7cbf-4b5a-ad36-f6cdabe27221-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.529956 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-krkl7"] Nov 28 15:45:16 crc kubenswrapper[4805]: E1128 15:45:16.530393 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b75302a-7cbf-4b5a-ad36-f6cdabe27221" containerName="glance-db-sync" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.530411 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b75302a-7cbf-4b5a-ad36-f6cdabe27221" containerName="glance-db-sync" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.530849 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b75302a-7cbf-4b5a-ad36-f6cdabe27221" containerName="glance-db-sync" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.531808 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.543383 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-krkl7"] Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.690030 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prp88\" (UniqueName: \"kubernetes.io/projected/a869fbea-46a1-4f70-891b-90a27a1527d1-kube-api-access-prp88\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.690324 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-nb\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.690352 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-dns-svc\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.690417 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-config\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.690449 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-sb\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.792316 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-nb\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.792395 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-dns-svc\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.792466 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-config\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.792507 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-sb\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.792560 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prp88\" (UniqueName: \"kubernetes.io/projected/a869fbea-46a1-4f70-891b-90a27a1527d1-kube-api-access-prp88\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.793878 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-nb\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.794498 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-config\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.794975 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-dns-svc\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.795168 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-sb\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.819667 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prp88\" (UniqueName: \"kubernetes.io/projected/a869fbea-46a1-4f70-891b-90a27a1527d1-kube-api-access-prp88\") pod \"dnsmasq-dns-6df78bdcfc-krkl7\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:16 crc kubenswrapper[4805]: I1128 15:45:16.856945 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:17 crc kubenswrapper[4805]: I1128 15:45:17.140222 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-krkl7"] Nov 28 15:45:17 crc kubenswrapper[4805]: W1128 15:45:17.145026 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda869fbea_46a1_4f70_891b_90a27a1527d1.slice/crio-7a750a29845f6cbd5a0ef6fa6210a5591035008c4dd28800edff9ac87839cb69 WatchSource:0}: Error finding container 7a750a29845f6cbd5a0ef6fa6210a5591035008c4dd28800edff9ac87839cb69: Status 404 returned error can't find the container with id 7a750a29845f6cbd5a0ef6fa6210a5591035008c4dd28800edff9ac87839cb69 Nov 28 15:45:17 crc kubenswrapper[4805]: I1128 15:45:17.193309 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" event={"ID":"a869fbea-46a1-4f70-891b-90a27a1527d1","Type":"ContainerStarted","Data":"7a750a29845f6cbd5a0ef6fa6210a5591035008c4dd28800edff9ac87839cb69"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.204630 4805 generic.go:334] "Generic (PLEG): container finished" podID="7a29bf56-ccc6-4b62-a4b7-da0093d80478" containerID="1d33f1bb0bf0611e99c60c4a3164ece57f7449f7387ff2182d439cdcab248b27" exitCode=0 Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.204663 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-1f9e-account-create-update-c5tc4" event={"ID":"7a29bf56-ccc6-4b62-a4b7-da0093d80478","Type":"ContainerDied","Data":"1d33f1bb0bf0611e99c60c4a3164ece57f7449f7387ff2182d439cdcab248b27"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.207976 4805 generic.go:334] "Generic (PLEG): container finished" podID="c308b8e2-19af-4cfc-9a06-487497e95198" containerID="5728fc21b8a8118968ef55e7ad974f51510ac290bf4297215efb9f590e045ec0" exitCode=0 Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.208005 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dnbhb" event={"ID":"c308b8e2-19af-4cfc-9a06-487497e95198","Type":"ContainerDied","Data":"5728fc21b8a8118968ef55e7ad974f51510ac290bf4297215efb9f590e045ec0"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.209761 4805 generic.go:334] "Generic (PLEG): container finished" podID="a869fbea-46a1-4f70-891b-90a27a1527d1" containerID="ebdb3b51aa00ed05f41aa993d8ac77042a1ba60f7c7205c1ecaf17706bd60c87" exitCode=0 Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.209818 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" event={"ID":"a869fbea-46a1-4f70-891b-90a27a1527d1","Type":"ContainerDied","Data":"ebdb3b51aa00ed05f41aa993d8ac77042a1ba60f7c7205c1ecaf17706bd60c87"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.211387 4805 generic.go:334] "Generic (PLEG): container finished" podID="09c6b57a-16a1-465d-b883-aa6bea6756c5" containerID="9043d7fdde1ce41cbf522437ac97e3b9791a64f0d7098912354924104f8f16f6" exitCode=0 Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.211517 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-72ad-account-create-update-66ckj" event={"ID":"09c6b57a-16a1-465d-b883-aa6bea6756c5","Type":"ContainerDied","Data":"9043d7fdde1ce41cbf522437ac97e3b9791a64f0d7098912354924104f8f16f6"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.241971 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"7413e61bbe43f8d00d9ac4ea2485eca35952ef72195b1eb40171e6705175ffcd"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.242021 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerStarted","Data":"841feae0d3f809ed9e79dff8941de56c70aafb0de8f991cd7453d48a9c67cc4c"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.249459 4805 generic.go:334] "Generic (PLEG): container finished" podID="19093324-6f6b-47d3-a08d-051d59136892" containerID="7a66ca4e8862507987fdb88cdd304c61afe27c97c176cd5e1cfb6c13d29f96ad" exitCode=0 Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.249690 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jfqsr" event={"ID":"19093324-6f6b-47d3-a08d-051d59136892","Type":"ContainerDied","Data":"7a66ca4e8862507987fdb88cdd304c61afe27c97c176cd5e1cfb6c13d29f96ad"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.253375 4805 generic.go:334] "Generic (PLEG): container finished" podID="ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce" containerID="6ac18cc12cbd5e678d45384978eaba73c92c3dd0e8f7dfc622bed25ab557e718" exitCode=0 Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.253437 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bzx4d" event={"ID":"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce","Type":"ContainerDied","Data":"6ac18cc12cbd5e678d45384978eaba73c92c3dd0e8f7dfc622bed25ab557e718"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.256057 4805 generic.go:334] "Generic (PLEG): container finished" podID="3c642d47-f1ea-4254-a76e-5ebe959da352" containerID="0ebb89ebcccc94757bb175b0c950e5a995e346c32fea6d70aa71be39c4c62c8d" exitCode=0 Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.257532 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8232-account-create-update-tg9lf" event={"ID":"3c642d47-f1ea-4254-a76e-5ebe959da352","Type":"ContainerDied","Data":"0ebb89ebcccc94757bb175b0c950e5a995e346c32fea6d70aa71be39c4c62c8d"} Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.367518 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=43.411244965 podStartE2EDuration="49.3674916s" podCreationTimestamp="2025-11-28 15:44:29 +0000 UTC" firstStartedPulling="2025-11-28 15:45:05.828741358 +0000 UTC m=+1132.878532669" lastFinishedPulling="2025-11-28 15:45:11.784987993 +0000 UTC m=+1138.834779304" observedRunningTime="2025-11-28 15:45:18.363939614 +0000 UTC m=+1145.413730925" watchObservedRunningTime="2025-11-28 15:45:18.3674916 +0000 UTC m=+1145.417282911" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.614152 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-krkl7"] Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.641159 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-9q7kt"] Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.642769 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.646854 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.658549 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-9q7kt"] Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.837967 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-swift-storage-0\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.838018 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.838085 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.838169 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-svc\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.838213 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmm5r\" (UniqueName: \"kubernetes.io/projected/bb986acb-fdca-4ba2-9289-fa37b9e06e62-kube-api-access-lmm5r\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.838243 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-config\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.939331 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.939665 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.939825 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-svc\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.939916 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmm5r\" (UniqueName: \"kubernetes.io/projected/bb986acb-fdca-4ba2-9289-fa37b9e06e62-kube-api-access-lmm5r\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.940002 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-config\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.940085 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-swift-storage-0\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.940185 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.940332 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.940703 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-svc\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.940976 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-config\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.941400 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-swift-storage-0\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.958774 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmm5r\" (UniqueName: \"kubernetes.io/projected/bb986acb-fdca-4ba2-9289-fa37b9e06e62-kube-api-access-lmm5r\") pod \"dnsmasq-dns-5bfc9d5487-9q7kt\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:18 crc kubenswrapper[4805]: I1128 15:45:18.978419 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:19 crc kubenswrapper[4805]: I1128 15:45:19.332744 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" event={"ID":"a869fbea-46a1-4f70-891b-90a27a1527d1","Type":"ContainerStarted","Data":"aa9513cfb6de7702a6b4bec85ab8f4e05e6b8e156f22d6759ee68c0788aa186f"} Nov 28 15:45:19 crc kubenswrapper[4805]: I1128 15:45:19.334156 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:19 crc kubenswrapper[4805]: I1128 15:45:19.371866 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" podStartSLOduration=3.371841503 podStartE2EDuration="3.371841503s" podCreationTimestamp="2025-11-28 15:45:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:45:19.36440983 +0000 UTC m=+1146.414201141" watchObservedRunningTime="2025-11-28 15:45:19.371841503 +0000 UTC m=+1146.421632824" Nov 28 15:45:19 crc kubenswrapper[4805]: I1128 15:45:19.494666 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-9q7kt"] Nov 28 15:45:20 crc kubenswrapper[4805]: I1128 15:45:20.341124 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" podUID="a869fbea-46a1-4f70-891b-90a27a1527d1" containerName="dnsmasq-dns" containerID="cri-o://aa9513cfb6de7702a6b4bec85ab8f4e05e6b8e156f22d6759ee68c0788aa186f" gracePeriod=10 Nov 28 15:45:21 crc kubenswrapper[4805]: I1128 15:45:21.350523 4805 generic.go:334] "Generic (PLEG): container finished" podID="a869fbea-46a1-4f70-891b-90a27a1527d1" containerID="aa9513cfb6de7702a6b4bec85ab8f4e05e6b8e156f22d6759ee68c0788aa186f" exitCode=0 Nov 28 15:45:21 crc kubenswrapper[4805]: I1128 15:45:21.350575 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" event={"ID":"a869fbea-46a1-4f70-891b-90a27a1527d1","Type":"ContainerDied","Data":"aa9513cfb6de7702a6b4bec85ab8f4e05e6b8e156f22d6759ee68c0788aa186f"} Nov 28 15:45:22 crc kubenswrapper[4805]: W1128 15:45:22.058726 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb986acb_fdca_4ba2_9289_fa37b9e06e62.slice/crio-86d8e612f8b9e90a3062a08ff9bf857097441dd066f236a0a3fd09305382218b WatchSource:0}: Error finding container 86d8e612f8b9e90a3062a08ff9bf857097441dd066f236a0a3fd09305382218b: Status 404 returned error can't find the container with id 86d8e612f8b9e90a3062a08ff9bf857097441dd066f236a0a3fd09305382218b Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.358865 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jfqsr" event={"ID":"19093324-6f6b-47d3-a08d-051d59136892","Type":"ContainerDied","Data":"9d047158c51e3bcb3f4c4a4091f7c4369f41e0387063e21a87dcfcc1102b6704"} Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.359109 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d047158c51e3bcb3f4c4a4091f7c4369f41e0387063e21a87dcfcc1102b6704" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.360161 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bzx4d" event={"ID":"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce","Type":"ContainerDied","Data":"7f5de6ea1047902dd8235ade08906548f8ace0bc4a666ce5c9c30355d4085975"} Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.360183 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f5de6ea1047902dd8235ade08906548f8ace0bc4a666ce5c9c30355d4085975" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.361384 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8232-account-create-update-tg9lf" event={"ID":"3c642d47-f1ea-4254-a76e-5ebe959da352","Type":"ContainerDied","Data":"0e6c01cb376d01229e815d582c24d17503b80982036fc45630f85e20467f0035"} Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.361434 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e6c01cb376d01229e815d582c24d17503b80982036fc45630f85e20467f0035" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.362929 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-1f9e-account-create-update-c5tc4" event={"ID":"7a29bf56-ccc6-4b62-a4b7-da0093d80478","Type":"ContainerDied","Data":"a9c4809489cbcdf21a54b37bb58a4528f6c3e7fb82b8b489d0d2c53019a4d184"} Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.362952 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9c4809489cbcdf21a54b37bb58a4528f6c3e7fb82b8b489d0d2c53019a4d184" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.364603 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" event={"ID":"bb986acb-fdca-4ba2-9289-fa37b9e06e62","Type":"ContainerStarted","Data":"86d8e612f8b9e90a3062a08ff9bf857097441dd066f236a0a3fd09305382218b"} Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.365845 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dnbhb" event={"ID":"c308b8e2-19af-4cfc-9a06-487497e95198","Type":"ContainerDied","Data":"a274179c43990cd11524eacb2c0b21565d486ccc99011fd110aad43d08e0f146"} Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.365873 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a274179c43990cd11524eacb2c0b21565d486ccc99011fd110aad43d08e0f146" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.367258 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" event={"ID":"a869fbea-46a1-4f70-891b-90a27a1527d1","Type":"ContainerDied","Data":"7a750a29845f6cbd5a0ef6fa6210a5591035008c4dd28800edff9ac87839cb69"} Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.367281 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a750a29845f6cbd5a0ef6fa6210a5591035008c4dd28800edff9ac87839cb69" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.368656 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-72ad-account-create-update-66ckj" event={"ID":"09c6b57a-16a1-465d-b883-aa6bea6756c5","Type":"ContainerDied","Data":"34e1bd466c219c7477ab72652d75dbcb00f7e2a2578eb875c677493abc6d67a4"} Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.368677 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34e1bd466c219c7477ab72652d75dbcb00f7e2a2578eb875c677493abc6d67a4" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.389532 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.411250 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s72b9\" (UniqueName: \"kubernetes.io/projected/19093324-6f6b-47d3-a08d-051d59136892-kube-api-access-s72b9\") pod \"19093324-6f6b-47d3-a08d-051d59136892\" (UID: \"19093324-6f6b-47d3-a08d-051d59136892\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.411308 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19093324-6f6b-47d3-a08d-051d59136892-operator-scripts\") pod \"19093324-6f6b-47d3-a08d-051d59136892\" (UID: \"19093324-6f6b-47d3-a08d-051d59136892\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.412524 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19093324-6f6b-47d3-a08d-051d59136892-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "19093324-6f6b-47d3-a08d-051d59136892" (UID: "19093324-6f6b-47d3-a08d-051d59136892"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.416764 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19093324-6f6b-47d3-a08d-051d59136892-kube-api-access-s72b9" (OuterVolumeSpecName: "kube-api-access-s72b9") pod "19093324-6f6b-47d3-a08d-051d59136892" (UID: "19093324-6f6b-47d3-a08d-051d59136892"). InnerVolumeSpecName "kube-api-access-s72b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.464540 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.477470 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.492350 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.507118 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.513274 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09c6b57a-16a1-465d-b883-aa6bea6756c5-operator-scripts\") pod \"09c6b57a-16a1-465d-b883-aa6bea6756c5\" (UID: \"09c6b57a-16a1-465d-b883-aa6bea6756c5\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.513384 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c308b8e2-19af-4cfc-9a06-487497e95198-operator-scripts\") pod \"c308b8e2-19af-4cfc-9a06-487497e95198\" (UID: \"c308b8e2-19af-4cfc-9a06-487497e95198\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.513453 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxqnw\" (UniqueName: \"kubernetes.io/projected/09c6b57a-16a1-465d-b883-aa6bea6756c5-kube-api-access-xxqnw\") pod \"09c6b57a-16a1-465d-b883-aa6bea6756c5\" (UID: \"09c6b57a-16a1-465d-b883-aa6bea6756c5\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.513551 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjd4f\" (UniqueName: \"kubernetes.io/projected/c308b8e2-19af-4cfc-9a06-487497e95198-kube-api-access-vjd4f\") pod \"c308b8e2-19af-4cfc-9a06-487497e95198\" (UID: \"c308b8e2-19af-4cfc-9a06-487497e95198\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.513595 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52942\" (UniqueName: \"kubernetes.io/projected/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-kube-api-access-52942\") pod \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\" (UID: \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.513668 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a29bf56-ccc6-4b62-a4b7-da0093d80478-operator-scripts\") pod \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\" (UID: \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.513715 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlbfr\" (UniqueName: \"kubernetes.io/projected/7a29bf56-ccc6-4b62-a4b7-da0093d80478-kube-api-access-zlbfr\") pod \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\" (UID: \"7a29bf56-ccc6-4b62-a4b7-da0093d80478\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.513737 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-operator-scripts\") pod \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\" (UID: \"ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.515222 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s72b9\" (UniqueName: \"kubernetes.io/projected/19093324-6f6b-47d3-a08d-051d59136892-kube-api-access-s72b9\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.515256 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19093324-6f6b-47d3-a08d-051d59136892-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.518313 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c308b8e2-19af-4cfc-9a06-487497e95198-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c308b8e2-19af-4cfc-9a06-487497e95198" (UID: "c308b8e2-19af-4cfc-9a06-487497e95198"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.518649 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09c6b57a-16a1-465d-b883-aa6bea6756c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "09c6b57a-16a1-465d-b883-aa6bea6756c5" (UID: "09c6b57a-16a1-465d-b883-aa6bea6756c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.518925 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a29bf56-ccc6-4b62-a4b7-da0093d80478-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7a29bf56-ccc6-4b62-a4b7-da0093d80478" (UID: "7a29bf56-ccc6-4b62-a4b7-da0093d80478"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.521912 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c308b8e2-19af-4cfc-9a06-487497e95198-kube-api-access-vjd4f" (OuterVolumeSpecName: "kube-api-access-vjd4f") pod "c308b8e2-19af-4cfc-9a06-487497e95198" (UID: "c308b8e2-19af-4cfc-9a06-487497e95198"). InnerVolumeSpecName "kube-api-access-vjd4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.523571 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.523574 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09c6b57a-16a1-465d-b883-aa6bea6756c5-kube-api-access-xxqnw" (OuterVolumeSpecName: "kube-api-access-xxqnw") pod "09c6b57a-16a1-465d-b883-aa6bea6756c5" (UID: "09c6b57a-16a1-465d-b883-aa6bea6756c5"). InnerVolumeSpecName "kube-api-access-xxqnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.524511 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-kube-api-access-52942" (OuterVolumeSpecName: "kube-api-access-52942") pod "ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce" (UID: "ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce"). InnerVolumeSpecName "kube-api-access-52942". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.524712 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce" (UID: "ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.525533 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a29bf56-ccc6-4b62-a4b7-da0093d80478-kube-api-access-zlbfr" (OuterVolumeSpecName: "kube-api-access-zlbfr") pod "7a29bf56-ccc6-4b62-a4b7-da0093d80478" (UID: "7a29bf56-ccc6-4b62-a4b7-da0093d80478"). InnerVolumeSpecName "kube-api-access-zlbfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.541106 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.619164 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-nb\") pod \"a869fbea-46a1-4f70-891b-90a27a1527d1\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.619623 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-dns-svc\") pod \"a869fbea-46a1-4f70-891b-90a27a1527d1\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.619712 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-sb\") pod \"a869fbea-46a1-4f70-891b-90a27a1527d1\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.619820 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgkd9\" (UniqueName: \"kubernetes.io/projected/3c642d47-f1ea-4254-a76e-5ebe959da352-kube-api-access-vgkd9\") pod \"3c642d47-f1ea-4254-a76e-5ebe959da352\" (UID: \"3c642d47-f1ea-4254-a76e-5ebe959da352\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.619966 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prp88\" (UniqueName: \"kubernetes.io/projected/a869fbea-46a1-4f70-891b-90a27a1527d1-kube-api-access-prp88\") pod \"a869fbea-46a1-4f70-891b-90a27a1527d1\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.619999 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-config\") pod \"a869fbea-46a1-4f70-891b-90a27a1527d1\" (UID: \"a869fbea-46a1-4f70-891b-90a27a1527d1\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.620054 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c642d47-f1ea-4254-a76e-5ebe959da352-operator-scripts\") pod \"3c642d47-f1ea-4254-a76e-5ebe959da352\" (UID: \"3c642d47-f1ea-4254-a76e-5ebe959da352\") " Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.620693 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a29bf56-ccc6-4b62-a4b7-da0093d80478-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.620714 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlbfr\" (UniqueName: \"kubernetes.io/projected/7a29bf56-ccc6-4b62-a4b7-da0093d80478-kube-api-access-zlbfr\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.620728 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.620739 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09c6b57a-16a1-465d-b883-aa6bea6756c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.620771 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c308b8e2-19af-4cfc-9a06-487497e95198-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.620783 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxqnw\" (UniqueName: \"kubernetes.io/projected/09c6b57a-16a1-465d-b883-aa6bea6756c5-kube-api-access-xxqnw\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.620794 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjd4f\" (UniqueName: \"kubernetes.io/projected/c308b8e2-19af-4cfc-9a06-487497e95198-kube-api-access-vjd4f\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.620804 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52942\" (UniqueName: \"kubernetes.io/projected/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce-kube-api-access-52942\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.621337 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c642d47-f1ea-4254-a76e-5ebe959da352-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c642d47-f1ea-4254-a76e-5ebe959da352" (UID: "3c642d47-f1ea-4254-a76e-5ebe959da352"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.639186 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a869fbea-46a1-4f70-891b-90a27a1527d1-kube-api-access-prp88" (OuterVolumeSpecName: "kube-api-access-prp88") pod "a869fbea-46a1-4f70-891b-90a27a1527d1" (UID: "a869fbea-46a1-4f70-891b-90a27a1527d1"). InnerVolumeSpecName "kube-api-access-prp88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.641669 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c642d47-f1ea-4254-a76e-5ebe959da352-kube-api-access-vgkd9" (OuterVolumeSpecName: "kube-api-access-vgkd9") pod "3c642d47-f1ea-4254-a76e-5ebe959da352" (UID: "3c642d47-f1ea-4254-a76e-5ebe959da352"). InnerVolumeSpecName "kube-api-access-vgkd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.663111 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a869fbea-46a1-4f70-891b-90a27a1527d1" (UID: "a869fbea-46a1-4f70-891b-90a27a1527d1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.666841 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a869fbea-46a1-4f70-891b-90a27a1527d1" (UID: "a869fbea-46a1-4f70-891b-90a27a1527d1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.671536 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a869fbea-46a1-4f70-891b-90a27a1527d1" (UID: "a869fbea-46a1-4f70-891b-90a27a1527d1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.677068 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-config" (OuterVolumeSpecName: "config") pod "a869fbea-46a1-4f70-891b-90a27a1527d1" (UID: "a869fbea-46a1-4f70-891b-90a27a1527d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.721601 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgkd9\" (UniqueName: \"kubernetes.io/projected/3c642d47-f1ea-4254-a76e-5ebe959da352-kube-api-access-vgkd9\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.721643 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prp88\" (UniqueName: \"kubernetes.io/projected/a869fbea-46a1-4f70-891b-90a27a1527d1-kube-api-access-prp88\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.721656 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.721669 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c642d47-f1ea-4254-a76e-5ebe959da352-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.721681 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.721693 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:22 crc kubenswrapper[4805]: I1128 15:45:22.721704 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a869fbea-46a1-4f70-891b-90a27a1527d1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.378270 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ct546" event={"ID":"7ac018eb-cb26-4d12-ac87-c200b20a9327","Type":"ContainerStarted","Data":"3aee0668fd5c39ac3a3e4f335457cb6946f12c017da091b68fe6eda24c00f29b"} Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.381159 4805 generic.go:334] "Generic (PLEG): container finished" podID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" containerID="13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8" exitCode=0 Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.381259 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dnbhb" Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.382227 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" event={"ID":"bb986acb-fdca-4ba2-9289-fa37b9e06e62","Type":"ContainerDied","Data":"13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8"} Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.382282 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72ad-account-create-update-66ckj" Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.382626 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bzx4d" Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.382932 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df78bdcfc-krkl7" Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.383263 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-1f9e-account-create-update-c5tc4" Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.383579 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jfqsr" Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.384308 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8232-account-create-update-tg9lf" Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.409287 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ct546" podStartSLOduration=2.411617177 podStartE2EDuration="9.409265362s" podCreationTimestamp="2025-11-28 15:45:14 +0000 UTC" firstStartedPulling="2025-11-28 15:45:15.134993496 +0000 UTC m=+1142.184784807" lastFinishedPulling="2025-11-28 15:45:22.132641681 +0000 UTC m=+1149.182432992" observedRunningTime="2025-11-28 15:45:23.402898689 +0000 UTC m=+1150.452690010" watchObservedRunningTime="2025-11-28 15:45:23.409265362 +0000 UTC m=+1150.459056683" Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.575368 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-krkl7"] Nov 28 15:45:23 crc kubenswrapper[4805]: I1128 15:45:23.590404 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-krkl7"] Nov 28 15:45:24 crc kubenswrapper[4805]: I1128 15:45:24.398936 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" event={"ID":"bb986acb-fdca-4ba2-9289-fa37b9e06e62","Type":"ContainerStarted","Data":"830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f"} Nov 28 15:45:24 crc kubenswrapper[4805]: I1128 15:45:24.399751 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:24 crc kubenswrapper[4805]: I1128 15:45:24.429135 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" podStartSLOduration=6.429114745 podStartE2EDuration="6.429114745s" podCreationTimestamp="2025-11-28 15:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:45:24.423334298 +0000 UTC m=+1151.473125609" watchObservedRunningTime="2025-11-28 15:45:24.429114745 +0000 UTC m=+1151.478906056" Nov 28 15:45:25 crc kubenswrapper[4805]: I1128 15:45:25.215451 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a869fbea-46a1-4f70-891b-90a27a1527d1" path="/var/lib/kubelet/pods/a869fbea-46a1-4f70-891b-90a27a1527d1/volumes" Nov 28 15:45:26 crc kubenswrapper[4805]: I1128 15:45:26.416579 4805 generic.go:334] "Generic (PLEG): container finished" podID="7ac018eb-cb26-4d12-ac87-c200b20a9327" containerID="3aee0668fd5c39ac3a3e4f335457cb6946f12c017da091b68fe6eda24c00f29b" exitCode=0 Nov 28 15:45:26 crc kubenswrapper[4805]: I1128 15:45:26.416676 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ct546" event={"ID":"7ac018eb-cb26-4d12-ac87-c200b20a9327","Type":"ContainerDied","Data":"3aee0668fd5c39ac3a3e4f335457cb6946f12c017da091b68fe6eda24c00f29b"} Nov 28 15:45:27 crc kubenswrapper[4805]: I1128 15:45:27.811374 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:27 crc kubenswrapper[4805]: I1128 15:45:27.909489 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-combined-ca-bundle\") pod \"7ac018eb-cb26-4d12-ac87-c200b20a9327\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " Nov 28 15:45:27 crc kubenswrapper[4805]: I1128 15:45:27.909579 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfw26\" (UniqueName: \"kubernetes.io/projected/7ac018eb-cb26-4d12-ac87-c200b20a9327-kube-api-access-gfw26\") pod \"7ac018eb-cb26-4d12-ac87-c200b20a9327\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " Nov 28 15:45:27 crc kubenswrapper[4805]: I1128 15:45:27.909905 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-config-data\") pod \"7ac018eb-cb26-4d12-ac87-c200b20a9327\" (UID: \"7ac018eb-cb26-4d12-ac87-c200b20a9327\") " Nov 28 15:45:27 crc kubenswrapper[4805]: I1128 15:45:27.915729 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac018eb-cb26-4d12-ac87-c200b20a9327-kube-api-access-gfw26" (OuterVolumeSpecName: "kube-api-access-gfw26") pod "7ac018eb-cb26-4d12-ac87-c200b20a9327" (UID: "7ac018eb-cb26-4d12-ac87-c200b20a9327"). InnerVolumeSpecName "kube-api-access-gfw26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:27 crc kubenswrapper[4805]: I1128 15:45:27.937780 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ac018eb-cb26-4d12-ac87-c200b20a9327" (UID: "7ac018eb-cb26-4d12-ac87-c200b20a9327"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:27 crc kubenswrapper[4805]: I1128 15:45:27.967645 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-config-data" (OuterVolumeSpecName: "config-data") pod "7ac018eb-cb26-4d12-ac87-c200b20a9327" (UID: "7ac018eb-cb26-4d12-ac87-c200b20a9327"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.011811 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfw26\" (UniqueName: \"kubernetes.io/projected/7ac018eb-cb26-4d12-ac87-c200b20a9327-kube-api-access-gfw26\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.012143 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.012157 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac018eb-cb26-4d12-ac87-c200b20a9327-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.457543 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ct546" event={"ID":"7ac018eb-cb26-4d12-ac87-c200b20a9327","Type":"ContainerDied","Data":"3c08b756ec5332e4fcb92d5c64d172647c637bd580632d09a5aa967347cef0f8"} Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.457584 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c08b756ec5332e4fcb92d5c64d172647c637bd580632d09a5aa967347cef0f8" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.457641 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ct546" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.599484 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-9q7kt"] Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.599689 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" podUID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" containerName="dnsmasq-dns" containerID="cri-o://830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f" gracePeriod=10 Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.603736 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.648603 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-prbpr"] Nov 28 15:45:28 crc kubenswrapper[4805]: E1128 15:45:28.648942 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19093324-6f6b-47d3-a08d-051d59136892" containerName="mariadb-database-create" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.648958 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="19093324-6f6b-47d3-a08d-051d59136892" containerName="mariadb-database-create" Nov 28 15:45:28 crc kubenswrapper[4805]: E1128 15:45:28.648975 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a869fbea-46a1-4f70-891b-90a27a1527d1" containerName="dnsmasq-dns" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.648983 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a869fbea-46a1-4f70-891b-90a27a1527d1" containerName="dnsmasq-dns" Nov 28 15:45:28 crc kubenswrapper[4805]: E1128 15:45:28.648991 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac018eb-cb26-4d12-ac87-c200b20a9327" containerName="keystone-db-sync" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.648997 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac018eb-cb26-4d12-ac87-c200b20a9327" containerName="keystone-db-sync" Nov 28 15:45:28 crc kubenswrapper[4805]: E1128 15:45:28.649007 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a29bf56-ccc6-4b62-a4b7-da0093d80478" containerName="mariadb-account-create-update" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649013 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a29bf56-ccc6-4b62-a4b7-da0093d80478" containerName="mariadb-account-create-update" Nov 28 15:45:28 crc kubenswrapper[4805]: E1128 15:45:28.649025 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c6b57a-16a1-465d-b883-aa6bea6756c5" containerName="mariadb-account-create-update" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649030 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c6b57a-16a1-465d-b883-aa6bea6756c5" containerName="mariadb-account-create-update" Nov 28 15:45:28 crc kubenswrapper[4805]: E1128 15:45:28.649041 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce" containerName="mariadb-database-create" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649047 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce" containerName="mariadb-database-create" Nov 28 15:45:28 crc kubenswrapper[4805]: E1128 15:45:28.649058 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a869fbea-46a1-4f70-891b-90a27a1527d1" containerName="init" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649063 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a869fbea-46a1-4f70-891b-90a27a1527d1" containerName="init" Nov 28 15:45:28 crc kubenswrapper[4805]: E1128 15:45:28.649072 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c642d47-f1ea-4254-a76e-5ebe959da352" containerName="mariadb-account-create-update" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649078 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c642d47-f1ea-4254-a76e-5ebe959da352" containerName="mariadb-account-create-update" Nov 28 15:45:28 crc kubenswrapper[4805]: E1128 15:45:28.649095 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c308b8e2-19af-4cfc-9a06-487497e95198" containerName="mariadb-database-create" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649101 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c308b8e2-19af-4cfc-9a06-487497e95198" containerName="mariadb-database-create" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649295 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c308b8e2-19af-4cfc-9a06-487497e95198" containerName="mariadb-database-create" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649311 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="19093324-6f6b-47d3-a08d-051d59136892" containerName="mariadb-database-create" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649322 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a869fbea-46a1-4f70-891b-90a27a1527d1" containerName="dnsmasq-dns" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649331 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c642d47-f1ea-4254-a76e-5ebe959da352" containerName="mariadb-account-create-update" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649342 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce" containerName="mariadb-database-create" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649351 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="09c6b57a-16a1-465d-b883-aa6bea6756c5" containerName="mariadb-account-create-update" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649374 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a29bf56-ccc6-4b62-a4b7-da0093d80478" containerName="mariadb-account-create-update" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.649384 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac018eb-cb26-4d12-ac87-c200b20a9327" containerName="keystone-db-sync" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.650201 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.665215 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fj49w"] Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.666314 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.669265 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.670076 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fzt55" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.670228 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.670415 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.673663 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.679511 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-prbpr"] Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.710932 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fj49w"] Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.725232 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-svc\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.725269 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-config-data\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.725290 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-nb\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.725325 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-fernet-keys\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.725339 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-sb\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.729305 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-scripts\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.729451 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-combined-ca-bundle\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.729554 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-config\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.729596 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrg67\" (UniqueName: \"kubernetes.io/projected/4322e7fd-5397-4d32-a24c-1c3b299171a5-kube-api-access-xrg67\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.729742 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-swift-storage-0\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.729768 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc5ft\" (UniqueName: \"kubernetes.io/projected/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-kube-api-access-kc5ft\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.729804 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-credential-keys\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.834655 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-svc\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.834983 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-config-data\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835005 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-nb\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835047 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-fernet-keys\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835063 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-sb\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835090 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-scripts\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835119 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-combined-ca-bundle\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835155 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-config\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835176 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrg67\" (UniqueName: \"kubernetes.io/projected/4322e7fd-5397-4d32-a24c-1c3b299171a5-kube-api-access-xrg67\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835233 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-swift-storage-0\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835258 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc5ft\" (UniqueName: \"kubernetes.io/projected/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-kube-api-access-kc5ft\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.835303 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-credential-keys\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.841066 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-svc\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.841426 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-config-data\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.842012 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-config\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.842424 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-nb\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.842850 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-swift-storage-0\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.843067 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-sb\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.853447 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-credential-keys\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.861798 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-scripts\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.869914 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-combined-ca-bundle\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.880690 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc5ft\" (UniqueName: \"kubernetes.io/projected/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-kube-api-access-kc5ft\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.881693 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-fernet-keys\") pod \"keystone-bootstrap-fj49w\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.884031 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrg67\" (UniqueName: \"kubernetes.io/projected/4322e7fd-5397-4d32-a24c-1c3b299171a5-kube-api-access-xrg67\") pod \"dnsmasq-dns-65c6dfc787-prbpr\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.895417 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-zwp6n"] Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.896566 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.904822 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.905083 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gddf4" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.905608 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.911974 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-zwp6n"] Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.936165 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bvd5\" (UniqueName: \"kubernetes.io/projected/b2f218fa-9d63-46f4-ba0d-3b9166d23672-kube-api-access-9bvd5\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.936280 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-scripts\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.936304 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-db-sync-config-data\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.936327 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-combined-ca-bundle\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.936455 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b2f218fa-9d63-46f4-ba0d-3b9166d23672-etc-machine-id\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.936474 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-config-data\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.944762 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.953332 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.958830 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.959039 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.981944 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" podUID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.984617 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-prbpr"] Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.985024 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:28 crc kubenswrapper[4805]: I1128 15:45:28.995274 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.038978 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-config-data\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039018 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039045 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039068 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-scripts\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039085 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-db-sync-config-data\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039109 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-combined-ca-bundle\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039132 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvv7w\" (UniqueName: \"kubernetes.io/projected/122b157a-374d-437e-97bb-66fa032893ff-kube-api-access-tvv7w\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039157 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-scripts\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039173 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-log-httpd\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039201 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b2f218fa-9d63-46f4-ba0d-3b9166d23672-etc-machine-id\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039219 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-config-data\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039274 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bvd5\" (UniqueName: \"kubernetes.io/projected/b2f218fa-9d63-46f4-ba0d-3b9166d23672-kube-api-access-9bvd5\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.039312 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-run-httpd\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.041462 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b2f218fa-9d63-46f4-ba0d-3b9166d23672-etc-machine-id\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.046694 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-scripts\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.047910 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-db-sync-config-data\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.065935 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-config-data\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.070197 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f645789c-np5d7"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.071606 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.081783 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-kfvps"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.088454 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.098620 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-combined-ca-bundle\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.098870 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gzh2w" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.099038 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.100315 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bvd5\" (UniqueName: \"kubernetes.io/projected/b2f218fa-9d63-46f4-ba0d-3b9166d23672-kube-api-access-9bvd5\") pod \"cinder-db-sync-zwp6n\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.118518 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-zr49k"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.119585 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.128723 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.132013 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dwvnp" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.132141 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.132229 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-kfvps"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.132557 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.140361 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-np5d7"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.140918 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-run-httpd\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.140963 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-config-data\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.140984 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-combined-ca-bundle\") pod \"barbican-db-sync-kfvps\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141018 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141042 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141075 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-config\") pod \"neutron-db-sync-zr49k\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141092 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvv7w\" (UniqueName: \"kubernetes.io/projected/122b157a-374d-437e-97bb-66fa032893ff-kube-api-access-tvv7w\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141116 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-scripts\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141131 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-log-httpd\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141157 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-combined-ca-bundle\") pod \"neutron-db-sync-zr49k\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141173 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-nb\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141201 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qczgq\" (UniqueName: \"kubernetes.io/projected/e013d7bd-1c72-4b87-be4b-83f895797124-kube-api-access-qczgq\") pod \"neutron-db-sync-zr49k\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141223 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhvtn\" (UniqueName: \"kubernetes.io/projected/ea1fc101-a095-4a2f-a0c3-9e13105a6077-kube-api-access-jhvtn\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141262 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-svc\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141277 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4c8q\" (UniqueName: \"kubernetes.io/projected/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-kube-api-access-f4c8q\") pod \"barbican-db-sync-kfvps\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141293 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-swift-storage-0\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141383 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-config\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141413 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-db-sync-config-data\") pod \"barbican-db-sync-kfvps\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.141448 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-sb\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.142034 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-run-httpd\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.144715 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-log-httpd\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.145782 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.161453 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-hbmtq"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.164847 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-config-data\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.165287 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.175838 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-l4wfm" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.176142 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.176293 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.176446 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.177394 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-zr49k"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.188711 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-scripts\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.194404 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hbmtq"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.198133 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvv7w\" (UniqueName: \"kubernetes.io/projected/122b157a-374d-437e-97bb-66fa032893ff-kube-api-access-tvv7w\") pod \"ceilometer-0\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.267784 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-svc\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.267854 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4c8q\" (UniqueName: \"kubernetes.io/projected/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-kube-api-access-f4c8q\") pod \"barbican-db-sync-kfvps\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.267874 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-swift-storage-0\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.267987 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-config\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.268050 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-db-sync-config-data\") pod \"barbican-db-sync-kfvps\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.268140 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-sb\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.268208 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-combined-ca-bundle\") pod \"barbican-db-sync-kfvps\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.268794 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-config\") pod \"neutron-db-sync-zr49k\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.268877 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-combined-ca-bundle\") pod \"neutron-db-sync-zr49k\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.268902 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-nb\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.268933 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qczgq\" (UniqueName: \"kubernetes.io/projected/e013d7bd-1c72-4b87-be4b-83f895797124-kube-api-access-qczgq\") pod \"neutron-db-sync-zr49k\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.268955 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhvtn\" (UniqueName: \"kubernetes.io/projected/ea1fc101-a095-4a2f-a0c3-9e13105a6077-kube-api-access-jhvtn\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.271130 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-nb\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.275129 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-config\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.275306 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-config\") pod \"neutron-db-sync-zr49k\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.277859 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-swift-storage-0\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.277979 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-sb\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.278622 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-svc\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.285874 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-combined-ca-bundle\") pod \"neutron-db-sync-zr49k\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.286222 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.290683 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qczgq\" (UniqueName: \"kubernetes.io/projected/e013d7bd-1c72-4b87-be4b-83f895797124-kube-api-access-qczgq\") pod \"neutron-db-sync-zr49k\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.291330 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhvtn\" (UniqueName: \"kubernetes.io/projected/ea1fc101-a095-4a2f-a0c3-9e13105a6077-kube-api-access-jhvtn\") pod \"dnsmasq-dns-55f645789c-np5d7\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.297771 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.304836 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.307152 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4c8q\" (UniqueName: \"kubernetes.io/projected/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-kube-api-access-f4c8q\") pod \"barbican-db-sync-kfvps\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.319328 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-combined-ca-bundle\") pod \"barbican-db-sync-kfvps\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.330326 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-db-sync-config-data\") pod \"barbican-db-sync-kfvps\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.342317 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zr49k" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.370740 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-scripts\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.370861 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37bb6d65-078d-4d19-bc10-dc50ea71edae-logs\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.370978 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-config-data\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.371194 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-combined-ca-bundle\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.371298 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sffpv\" (UniqueName: \"kubernetes.io/projected/37bb6d65-078d-4d19-bc10-dc50ea71edae-kube-api-access-sffpv\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.378775 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.473282 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-scripts\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.473347 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37bb6d65-078d-4d19-bc10-dc50ea71edae-logs\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.473401 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-config-data\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.473469 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-combined-ca-bundle\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.473523 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sffpv\" (UniqueName: \"kubernetes.io/projected/37bb6d65-078d-4d19-bc10-dc50ea71edae-kube-api-access-sffpv\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.474070 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37bb6d65-078d-4d19-bc10-dc50ea71edae-logs\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.482561 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-combined-ca-bundle\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.490850 4805 generic.go:334] "Generic (PLEG): container finished" podID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" containerID="830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f" exitCode=0 Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.490896 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" event={"ID":"bb986acb-fdca-4ba2-9289-fa37b9e06e62","Type":"ContainerDied","Data":"830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f"} Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.490926 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" event={"ID":"bb986acb-fdca-4ba2-9289-fa37b9e06e62","Type":"ContainerDied","Data":"86d8e612f8b9e90a3062a08ff9bf857097441dd066f236a0a3fd09305382218b"} Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.490944 4805 scope.go:117] "RemoveContainer" containerID="830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.491073 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfc9d5487-9q7kt" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.496106 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sffpv\" (UniqueName: \"kubernetes.io/projected/37bb6d65-078d-4d19-bc10-dc50ea71edae-kube-api-access-sffpv\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.516546 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-scripts\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.516910 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-config-data\") pod \"placement-db-sync-hbmtq\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.563706 4805 scope.go:117] "RemoveContainer" containerID="13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.573932 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-sb\") pod \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.574026 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-svc\") pod \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.574126 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-nb\") pod \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.574209 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmm5r\" (UniqueName: \"kubernetes.io/projected/bb986acb-fdca-4ba2-9289-fa37b9e06e62-kube-api-access-lmm5r\") pod \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.574436 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-swift-storage-0\") pod \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.574512 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-config\") pod \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\" (UID: \"bb986acb-fdca-4ba2-9289-fa37b9e06e62\") " Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.591018 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb986acb-fdca-4ba2-9289-fa37b9e06e62-kube-api-access-lmm5r" (OuterVolumeSpecName: "kube-api-access-lmm5r") pod "bb986acb-fdca-4ba2-9289-fa37b9e06e62" (UID: "bb986acb-fdca-4ba2-9289-fa37b9e06e62"). InnerVolumeSpecName "kube-api-access-lmm5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.621274 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-kfvps" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.650316 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fj49w"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.655128 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hbmtq" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.676320 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmm5r\" (UniqueName: \"kubernetes.io/projected/bb986acb-fdca-4ba2-9289-fa37b9e06e62-kube-api-access-lmm5r\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.681731 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb986acb-fdca-4ba2-9289-fa37b9e06e62" (UID: "bb986acb-fdca-4ba2-9289-fa37b9e06e62"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.681998 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-prbpr"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.682519 4805 scope.go:117] "RemoveContainer" containerID="830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f" Nov 28 15:45:29 crc kubenswrapper[4805]: E1128 15:45:29.689611 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f\": container with ID starting with 830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f not found: ID does not exist" containerID="830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.689693 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f"} err="failed to get container status \"830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f\": rpc error: code = NotFound desc = could not find container \"830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f\": container with ID starting with 830bd1a90dea3a1d278ccba1ecce0dfe97b3d8500442f8f9fbf6cfc6d9885e8f not found: ID does not exist" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.689725 4805 scope.go:117] "RemoveContainer" containerID="13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8" Nov 28 15:45:29 crc kubenswrapper[4805]: E1128 15:45:29.691416 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8\": container with ID starting with 13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8 not found: ID does not exist" containerID="13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.691434 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8"} err="failed to get container status \"13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8\": rpc error: code = NotFound desc = could not find container \"13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8\": container with ID starting with 13a89f65b0b6b70bab613f0946edbcc6d025dff5b2087606bb600516d682d8c8 not found: ID does not exist" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.693909 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb986acb-fdca-4ba2-9289-fa37b9e06e62" (UID: "bb986acb-fdca-4ba2-9289-fa37b9e06e62"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.750529 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-config" (OuterVolumeSpecName: "config") pod "bb986acb-fdca-4ba2-9289-fa37b9e06e62" (UID: "bb986acb-fdca-4ba2-9289-fa37b9e06e62"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.756021 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:45:29 crc kubenswrapper[4805]: E1128 15:45:29.756474 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" containerName="dnsmasq-dns" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.756488 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" containerName="dnsmasq-dns" Nov 28 15:45:29 crc kubenswrapper[4805]: E1128 15:45:29.756835 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" containerName="init" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.756848 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" containerName="init" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.757552 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" containerName="dnsmasq-dns" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.758842 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.762761 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.762792 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-tjgr4" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.763052 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.763242 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.767233 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb986acb-fdca-4ba2-9289-fa37b9e06e62" (UID: "bb986acb-fdca-4ba2-9289-fa37b9e06e62"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.789040 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.789379 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.789390 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.789399 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.789102 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.808542 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bb986acb-fdca-4ba2-9289-fa37b9e06e62" (UID: "bb986acb-fdca-4ba2-9289-fa37b9e06e62"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.857703 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.859446 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.863905 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.865094 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.866430 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.893461 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.893743 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-scripts\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.894103 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.894220 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zfg6\" (UniqueName: \"kubernetes.io/projected/5af67a46-cd69-44ed-9b8f-b01e10b9452f-kube-api-access-7zfg6\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.894327 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.894445 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.894679 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-logs\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.894759 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-config-data\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.894877 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb986acb-fdca-4ba2-9289-fa37b9e06e62-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997360 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997457 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997495 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zfg6\" (UniqueName: \"kubernetes.io/projected/5af67a46-cd69-44ed-9b8f-b01e10b9452f-kube-api-access-7zfg6\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997526 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997559 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997609 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997635 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997664 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997698 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-logs\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997720 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997746 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-config-data\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997781 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997802 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-scripts\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997829 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997844 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.997865 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxr56\" (UniqueName: \"kubernetes.io/projected/ef18d029-97e5-4c6b-b93e-0a40f55816a2-kube-api-access-wxr56\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.998789 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-logs\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.998911 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 28 15:45:29 crc kubenswrapper[4805]: I1128 15:45:29.999847 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.003277 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.004449 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-scripts\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.007241 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.008201 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-config-data\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.035827 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zfg6\" (UniqueName: \"kubernetes.io/projected/5af67a46-cd69-44ed-9b8f-b01e10b9452f-kube-api-access-7zfg6\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.060918 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.068788 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.100188 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.100259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.100430 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.100457 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.100491 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxr56\" (UniqueName: \"kubernetes.io/projected/ef18d029-97e5-4c6b-b93e-0a40f55816a2-kube-api-access-wxr56\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.100565 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.100609 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.100662 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.102023 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.102715 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.104503 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.108560 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.111719 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.111749 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.112600 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.127704 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxr56\" (UniqueName: \"kubernetes.io/projected/ef18d029-97e5-4c6b-b93e-0a40f55816a2-kube-api-access-wxr56\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.163994 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-9q7kt"] Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.177749 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.193040 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-9q7kt"] Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.306439 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-zwp6n"] Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.337468 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-zr49k"] Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.398479 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.446703 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-np5d7"] Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.451471 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.458711 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hbmtq"] Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.522438 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hbmtq" event={"ID":"37bb6d65-078d-4d19-bc10-dc50ea71edae","Type":"ContainerStarted","Data":"fcfc234ef249807518798b35eac6dcf6ff04a0eead585c05895e89101d2c4432"} Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.524260 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" event={"ID":"4322e7fd-5397-4d32-a24c-1c3b299171a5","Type":"ContainerStarted","Data":"7e14142f115840de9248d63a658c1ac834c8c89f4dc64664570ecab53be07ce8"} Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.525332 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-np5d7" event={"ID":"ea1fc101-a095-4a2f-a0c3-9e13105a6077","Type":"ContainerStarted","Data":"0df66dd49ed31efd381ea52d0e6e91bd79b8cf082aba66f5095a3a8dcfccb00c"} Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.526266 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fj49w" event={"ID":"abb57ada-d9f9-4162-a8f5-dd3f88a22f50","Type":"ContainerStarted","Data":"bb14b64653965cdb12040c9c92f15b133dca4eb5cdc076c177bea0a664a8e0bd"} Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.527213 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zr49k" event={"ID":"e013d7bd-1c72-4b87-be4b-83f895797124","Type":"ContainerStarted","Data":"cd7ca3cb627038937bf5091115d44b0df30fd12144e13d061d3e5c763fd3a333"} Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.528625 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zwp6n" event={"ID":"b2f218fa-9d63-46f4-ba0d-3b9166d23672","Type":"ContainerStarted","Data":"f2a51e83a816cbb3a6cc732bf2fc8ded675b42b1c3b3cb3176308ccc3ab2c1f4"} Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.529610 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"122b157a-374d-437e-97bb-66fa032893ff","Type":"ContainerStarted","Data":"bb341155e49ee01dfea40123ea6e5bad22b83e1cc82b5f1492d7f6cfb18dfd41"} Nov 28 15:45:30 crc kubenswrapper[4805]: I1128 15:45:30.560913 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-kfvps"] Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:31.216906 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb986acb-fdca-4ba2-9289-fa37b9e06e62" path="/var/lib/kubelet/pods/bb986acb-fdca-4ba2-9289-fa37b9e06e62/volumes" Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:31.539752 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-kfvps" event={"ID":"e63a3a44-329f-4d06-af10-a8ac7c72dcc4","Type":"ContainerStarted","Data":"a14d597375e8964a17ca4d513211ec5b60b710e3572322748d5b95c1f34d5673"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:31.570034 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:45:36 crc kubenswrapper[4805]: W1128 15:45:31.574900 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5af67a46_cd69_44ed_9b8f_b01e10b9452f.slice/crio-3d441bcbd70c7cf28b125bbcdb74ebf85fe5f48caa7e44d1dfb4e7ac475b20ce WatchSource:0}: Error finding container 3d441bcbd70c7cf28b125bbcdb74ebf85fe5f48caa7e44d1dfb4e7ac475b20ce: Status 404 returned error can't find the container with id 3d441bcbd70c7cf28b125bbcdb74ebf85fe5f48caa7e44d1dfb4e7ac475b20ce Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:31.686659 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:31.758524 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:45:36 crc kubenswrapper[4805]: W1128 15:45:32.310814 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef18d029_97e5_4c6b_b93e_0a40f55816a2.slice/crio-3d9a85980d71759f88a5ed4cb21763aea8c50c82bc1b3c452cdd0a35cfbe3cf5 WatchSource:0}: Error finding container 3d9a85980d71759f88a5ed4cb21763aea8c50c82bc1b3c452cdd0a35cfbe3cf5: Status 404 returned error can't find the container with id 3d9a85980d71759f88a5ed4cb21763aea8c50c82bc1b3c452cdd0a35cfbe3cf5 Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:32.315728 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:32.550897 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ef18d029-97e5-4c6b-b93e-0a40f55816a2","Type":"ContainerStarted","Data":"3d9a85980d71759f88a5ed4cb21763aea8c50c82bc1b3c452cdd0a35cfbe3cf5"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:32.568451 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" event={"ID":"4322e7fd-5397-4d32-a24c-1c3b299171a5","Type":"ContainerStarted","Data":"96ccfd45f61b250ae5d9a05706495e9330c8aa06717bc09526eff352434d44bf"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:32.572065 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5af67a46-cd69-44ed-9b8f-b01e10b9452f","Type":"ContainerStarted","Data":"3d441bcbd70c7cf28b125bbcdb74ebf85fe5f48caa7e44d1dfb4e7ac475b20ce"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.582610 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ef18d029-97e5-4c6b-b93e-0a40f55816a2","Type":"ContainerStarted","Data":"5a22dde5461cfab341af8734383933aea39409c34b8df382b72cc7579b8593fe"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.584434 4805 generic.go:334] "Generic (PLEG): container finished" podID="4322e7fd-5397-4d32-a24c-1c3b299171a5" containerID="96ccfd45f61b250ae5d9a05706495e9330c8aa06717bc09526eff352434d44bf" exitCode=0 Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.584489 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" event={"ID":"4322e7fd-5397-4d32-a24c-1c3b299171a5","Type":"ContainerDied","Data":"96ccfd45f61b250ae5d9a05706495e9330c8aa06717bc09526eff352434d44bf"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.587987 4805 generic.go:334] "Generic (PLEG): container finished" podID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerID="138024a909782c1f5bbc8a05dcb191e646e71777eca84bbe32c86ae23a29abab" exitCode=0 Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.588048 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-np5d7" event={"ID":"ea1fc101-a095-4a2f-a0c3-9e13105a6077","Type":"ContainerDied","Data":"138024a909782c1f5bbc8a05dcb191e646e71777eca84bbe32c86ae23a29abab"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.591166 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5af67a46-cd69-44ed-9b8f-b01e10b9452f","Type":"ContainerStarted","Data":"4d31996196b214e437cca615ee49eda8e1c92ff83d7c008e831d988afd9c2e9c"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.591210 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5af67a46-cd69-44ed-9b8f-b01e10b9452f","Type":"ContainerStarted","Data":"8f6f0afc837b022cd04e980a4e338bee80b78148bdf3b45649b526488e1ea13c"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.591246 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerName="glance-log" containerID="cri-o://8f6f0afc837b022cd04e980a4e338bee80b78148bdf3b45649b526488e1ea13c" gracePeriod=30 Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.591335 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerName="glance-httpd" containerID="cri-o://4d31996196b214e437cca615ee49eda8e1c92ff83d7c008e831d988afd9c2e9c" gracePeriod=30 Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.601517 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fj49w" event={"ID":"abb57ada-d9f9-4162-a8f5-dd3f88a22f50","Type":"ContainerStarted","Data":"4e4402e0fc9b295b789402eb980861bc790e80c1e7f03bfe24e34dbc5607e0b8"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.611296 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zr49k" event={"ID":"e013d7bd-1c72-4b87-be4b-83f895797124","Type":"ContainerStarted","Data":"7f4ba1f787d29d980bc0c209ec06ceece827e2ac14219237aa842b0f992266a1"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.641951 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.64192698 podStartE2EDuration="5.64192698s" podCreationTimestamp="2025-11-28 15:45:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:45:33.631160207 +0000 UTC m=+1160.680951538" watchObservedRunningTime="2025-11-28 15:45:33.64192698 +0000 UTC m=+1160.691718321" Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.659020 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fj49w" podStartSLOduration=5.658999855 podStartE2EDuration="5.658999855s" podCreationTimestamp="2025-11-28 15:45:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:45:33.647856872 +0000 UTC m=+1160.697648193" watchObservedRunningTime="2025-11-28 15:45:33.658999855 +0000 UTC m=+1160.708791166" Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:33.697388 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-zr49k" podStartSLOduration=4.697347388 podStartE2EDuration="4.697347388s" podCreationTimestamp="2025-11-28 15:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:45:33.689450634 +0000 UTC m=+1160.739241955" watchObservedRunningTime="2025-11-28 15:45:33.697347388 +0000 UTC m=+1160.747138699" Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:34.622011 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ef18d029-97e5-4c6b-b93e-0a40f55816a2","Type":"ContainerStarted","Data":"3bdc559c6aa17f08eca4f8a4b86e51ca90fd3eb7ad664f71ba07dd24f99a07d6"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:34.624623 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-np5d7" event={"ID":"ea1fc101-a095-4a2f-a0c3-9e13105a6077","Type":"ContainerStarted","Data":"851438c9ffd66c2d9c341bb6091b3cacc94b01f46299b3677347a2c85751bc4b"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:34.639567 4805 generic.go:334] "Generic (PLEG): container finished" podID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerID="4d31996196b214e437cca615ee49eda8e1c92ff83d7c008e831d988afd9c2e9c" exitCode=0 Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:34.639596 4805 generic.go:334] "Generic (PLEG): container finished" podID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerID="8f6f0afc837b022cd04e980a4e338bee80b78148bdf3b45649b526488e1ea13c" exitCode=143 Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:34.640755 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5af67a46-cd69-44ed-9b8f-b01e10b9452f","Type":"ContainerDied","Data":"4d31996196b214e437cca615ee49eda8e1c92ff83d7c008e831d988afd9c2e9c"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:34.640780 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5af67a46-cd69-44ed-9b8f-b01e10b9452f","Type":"ContainerDied","Data":"8f6f0afc837b022cd04e980a4e338bee80b78148bdf3b45649b526488e1ea13c"} Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:36.655809 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerName="glance-log" containerID="cri-o://5a22dde5461cfab341af8734383933aea39409c34b8df382b72cc7579b8593fe" gracePeriod=30 Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:36.655983 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerName="glance-httpd" containerID="cri-o://3bdc559c6aa17f08eca4f8a4b86e51ca90fd3eb7ad664f71ba07dd24f99a07d6" gracePeriod=30 Nov 28 15:45:36 crc kubenswrapper[4805]: I1128 15:45:36.695206 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.695187317 podStartE2EDuration="8.695187317s" podCreationTimestamp="2025-11-28 15:45:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:45:36.675340057 +0000 UTC m=+1163.725131358" watchObservedRunningTime="2025-11-28 15:45:36.695187317 +0000 UTC m=+1163.744978628" Nov 28 15:45:37 crc kubenswrapper[4805]: I1128 15:45:37.320792 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:45:37 crc kubenswrapper[4805]: I1128 15:45:37.668777 4805 generic.go:334] "Generic (PLEG): container finished" podID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerID="3bdc559c6aa17f08eca4f8a4b86e51ca90fd3eb7ad664f71ba07dd24f99a07d6" exitCode=0 Nov 28 15:45:37 crc kubenswrapper[4805]: I1128 15:45:37.669081 4805 generic.go:334] "Generic (PLEG): container finished" podID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerID="5a22dde5461cfab341af8734383933aea39409c34b8df382b72cc7579b8593fe" exitCode=143 Nov 28 15:45:37 crc kubenswrapper[4805]: I1128 15:45:37.670265 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ef18d029-97e5-4c6b-b93e-0a40f55816a2","Type":"ContainerDied","Data":"3bdc559c6aa17f08eca4f8a4b86e51ca90fd3eb7ad664f71ba07dd24f99a07d6"} Nov 28 15:45:37 crc kubenswrapper[4805]: I1128 15:45:37.670298 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ef18d029-97e5-4c6b-b93e-0a40f55816a2","Type":"ContainerDied","Data":"5a22dde5461cfab341af8734383933aea39409c34b8df382b72cc7579b8593fe"} Nov 28 15:45:37 crc kubenswrapper[4805]: I1128 15:45:37.670338 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:37 crc kubenswrapper[4805]: I1128 15:45:37.704248 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f645789c-np5d7" podStartSLOduration=8.704222076 podStartE2EDuration="8.704222076s" podCreationTimestamp="2025-11-28 15:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:45:37.700262618 +0000 UTC m=+1164.750053939" watchObservedRunningTime="2025-11-28 15:45:37.704222076 +0000 UTC m=+1164.754013387" Nov 28 15:45:39 crc kubenswrapper[4805]: I1128 15:45:39.309271 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:45:39 crc kubenswrapper[4805]: I1128 15:45:39.373561 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-4lt76"] Nov 28 15:45:39 crc kubenswrapper[4805]: I1128 15:45:39.373823 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="dnsmasq-dns" containerID="cri-o://05de3845991eaa359e6baa626dc02171b7f292526b958ada1028d01c0e1669d6" gracePeriod=10 Nov 28 15:45:39 crc kubenswrapper[4805]: I1128 15:45:39.756475 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.071858 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.133966 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-scripts\") pod \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.134021 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-logs\") pod \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.134047 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-combined-ca-bundle\") pod \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.134114 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-config-data\") pod \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.134232 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.134260 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-httpd-run\") pod \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.134343 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-public-tls-certs\") pod \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.134403 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zfg6\" (UniqueName: \"kubernetes.io/projected/5af67a46-cd69-44ed-9b8f-b01e10b9452f-kube-api-access-7zfg6\") pod \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\" (UID: \"5af67a46-cd69-44ed-9b8f-b01e10b9452f\") " Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.135761 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5af67a46-cd69-44ed-9b8f-b01e10b9452f" (UID: "5af67a46-cd69-44ed-9b8f-b01e10b9452f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.136021 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-logs" (OuterVolumeSpecName: "logs") pod "5af67a46-cd69-44ed-9b8f-b01e10b9452f" (UID: "5af67a46-cd69-44ed-9b8f-b01e10b9452f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.142508 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-scripts" (OuterVolumeSpecName: "scripts") pod "5af67a46-cd69-44ed-9b8f-b01e10b9452f" (UID: "5af67a46-cd69-44ed-9b8f-b01e10b9452f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.142509 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af67a46-cd69-44ed-9b8f-b01e10b9452f-kube-api-access-7zfg6" (OuterVolumeSpecName: "kube-api-access-7zfg6") pod "5af67a46-cd69-44ed-9b8f-b01e10b9452f" (UID: "5af67a46-cd69-44ed-9b8f-b01e10b9452f"). InnerVolumeSpecName "kube-api-access-7zfg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.158286 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "5af67a46-cd69-44ed-9b8f-b01e10b9452f" (UID: "5af67a46-cd69-44ed-9b8f-b01e10b9452f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.196651 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5af67a46-cd69-44ed-9b8f-b01e10b9452f" (UID: "5af67a46-cd69-44ed-9b8f-b01e10b9452f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.209785 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5af67a46-cd69-44ed-9b8f-b01e10b9452f" (UID: "5af67a46-cd69-44ed-9b8f-b01e10b9452f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.237699 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.237738 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.237752 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.237767 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zfg6\" (UniqueName: \"kubernetes.io/projected/5af67a46-cd69-44ed-9b8f-b01e10b9452f-kube-api-access-7zfg6\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.237778 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.237789 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5af67a46-cd69-44ed-9b8f-b01e10b9452f-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.237799 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.240471 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-config-data" (OuterVolumeSpecName: "config-data") pod "5af67a46-cd69-44ed-9b8f-b01e10b9452f" (UID: "5af67a46-cd69-44ed-9b8f-b01e10b9452f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.261873 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.339439 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af67a46-cd69-44ed-9b8f-b01e10b9452f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.339473 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.698006 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5af67a46-cd69-44ed-9b8f-b01e10b9452f","Type":"ContainerDied","Data":"3d441bcbd70c7cf28b125bbcdb74ebf85fe5f48caa7e44d1dfb4e7ac475b20ce"} Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.698086 4805 scope.go:117] "RemoveContainer" containerID="4d31996196b214e437cca615ee49eda8e1c92ff83d7c008e831d988afd9c2e9c" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.698092 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.702066 4805 generic.go:334] "Generic (PLEG): container finished" podID="abb57ada-d9f9-4162-a8f5-dd3f88a22f50" containerID="4e4402e0fc9b295b789402eb980861bc790e80c1e7f03bfe24e34dbc5607e0b8" exitCode=0 Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.702113 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fj49w" event={"ID":"abb57ada-d9f9-4162-a8f5-dd3f88a22f50","Type":"ContainerDied","Data":"4e4402e0fc9b295b789402eb980861bc790e80c1e7f03bfe24e34dbc5607e0b8"} Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.704470 4805 generic.go:334] "Generic (PLEG): container finished" podID="15970fdb-4e60-4331-bd19-40db152c2303" containerID="05de3845991eaa359e6baa626dc02171b7f292526b958ada1028d01c0e1669d6" exitCode=0 Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.704497 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" event={"ID":"15970fdb-4e60-4331-bd19-40db152c2303","Type":"ContainerDied","Data":"05de3845991eaa359e6baa626dc02171b7f292526b958ada1028d01c0e1669d6"} Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.739459 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.744516 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.774410 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:45:40 crc kubenswrapper[4805]: E1128 15:45:40.774821 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerName="glance-httpd" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.774836 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerName="glance-httpd" Nov 28 15:45:40 crc kubenswrapper[4805]: E1128 15:45:40.774858 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerName="glance-log" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.774866 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerName="glance-log" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.775078 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerName="glance-log" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.775103 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" containerName="glance-httpd" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.776133 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.780215 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.780394 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.782829 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.947984 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.948491 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.948558 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z2kz\" (UniqueName: \"kubernetes.io/projected/17e2c4d9-4960-4034-be79-5dffbab66c5b-kube-api-access-7z2kz\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.948607 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-config-data\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.948676 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.948699 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.948719 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-scripts\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:40 crc kubenswrapper[4805]: I1128 15:45:40.948746 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-logs\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.049845 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-config-data\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.049926 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.049944 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.049963 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-scripts\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.049983 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-logs\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.050027 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.050050 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.050090 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z2kz\" (UniqueName: \"kubernetes.io/projected/17e2c4d9-4960-4034-be79-5dffbab66c5b-kube-api-access-7z2kz\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.050708 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.051728 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-logs\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.051931 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.055765 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-config-data\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.058082 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.062256 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.069575 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z2kz\" (UniqueName: \"kubernetes.io/projected/17e2c4d9-4960-4034-be79-5dffbab66c5b-kube-api-access-7z2kz\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.070091 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-scripts\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.082903 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.124779 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:45:41 crc kubenswrapper[4805]: I1128 15:45:41.214263 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af67a46-cd69-44ed-9b8f-b01e10b9452f" path="/var/lib/kubelet/pods/5af67a46-cd69-44ed-9b8f-b01e10b9452f/volumes" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.033487 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.206448 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrg67\" (UniqueName: \"kubernetes.io/projected/4322e7fd-5397-4d32-a24c-1c3b299171a5-kube-api-access-xrg67\") pod \"4322e7fd-5397-4d32-a24c-1c3b299171a5\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.206495 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-sb\") pod \"4322e7fd-5397-4d32-a24c-1c3b299171a5\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.206525 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-nb\") pod \"4322e7fd-5397-4d32-a24c-1c3b299171a5\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.206663 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-svc\") pod \"4322e7fd-5397-4d32-a24c-1c3b299171a5\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.206704 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-swift-storage-0\") pod \"4322e7fd-5397-4d32-a24c-1c3b299171a5\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.206721 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-config\") pod \"4322e7fd-5397-4d32-a24c-1c3b299171a5\" (UID: \"4322e7fd-5397-4d32-a24c-1c3b299171a5\") " Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.228799 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4322e7fd-5397-4d32-a24c-1c3b299171a5-kube-api-access-xrg67" (OuterVolumeSpecName: "kube-api-access-xrg67") pod "4322e7fd-5397-4d32-a24c-1c3b299171a5" (UID: "4322e7fd-5397-4d32-a24c-1c3b299171a5"). InnerVolumeSpecName "kube-api-access-xrg67". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.232470 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-config" (OuterVolumeSpecName: "config") pod "4322e7fd-5397-4d32-a24c-1c3b299171a5" (UID: "4322e7fd-5397-4d32-a24c-1c3b299171a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.234297 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4322e7fd-5397-4d32-a24c-1c3b299171a5" (UID: "4322e7fd-5397-4d32-a24c-1c3b299171a5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.234807 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4322e7fd-5397-4d32-a24c-1c3b299171a5" (UID: "4322e7fd-5397-4d32-a24c-1c3b299171a5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.234849 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4322e7fd-5397-4d32-a24c-1c3b299171a5" (UID: "4322e7fd-5397-4d32-a24c-1c3b299171a5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.242919 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4322e7fd-5397-4d32-a24c-1c3b299171a5" (UID: "4322e7fd-5397-4d32-a24c-1c3b299171a5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.308477 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.308513 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrg67\" (UniqueName: \"kubernetes.io/projected/4322e7fd-5397-4d32-a24c-1c3b299171a5-kube-api-access-xrg67\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.308530 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.308542 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.308553 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.308594 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4322e7fd-5397-4d32-a24c-1c3b299171a5-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.760115 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" event={"ID":"4322e7fd-5397-4d32-a24c-1c3b299171a5","Type":"ContainerDied","Data":"7e14142f115840de9248d63a658c1ac834c8c89f4dc64664570ecab53be07ce8"} Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.760179 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c6dfc787-prbpr" Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.823514 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-prbpr"] Nov 28 15:45:44 crc kubenswrapper[4805]: I1128 15:45:44.833424 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-prbpr"] Nov 28 15:45:45 crc kubenswrapper[4805]: I1128 15:45:45.218018 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4322e7fd-5397-4d32-a24c-1c3b299171a5" path="/var/lib/kubelet/pods/4322e7fd-5397-4d32-a24c-1c3b299171a5/volumes" Nov 28 15:45:49 crc kubenswrapper[4805]: I1128 15:45:49.756722 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Nov 28 15:45:54 crc kubenswrapper[4805]: I1128 15:45:54.758055 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Nov 28 15:45:54 crc kubenswrapper[4805]: I1128 15:45:54.758807 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.234072 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.243485 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.256459 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.343966 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-config\") pod \"15970fdb-4e60-4331-bd19-40db152c2303\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344025 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-logs\") pod \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344063 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc5ft\" (UniqueName: \"kubernetes.io/projected/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-kube-api-access-kc5ft\") pod \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344082 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxr56\" (UniqueName: \"kubernetes.io/projected/ef18d029-97e5-4c6b-b93e-0a40f55816a2-kube-api-access-wxr56\") pod \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344146 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-scripts\") pod \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344169 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344185 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-sb\") pod \"15970fdb-4e60-4331-bd19-40db152c2303\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344208 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-internal-tls-certs\") pod \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344241 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-httpd-run\") pod \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344259 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-credential-keys\") pod \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344276 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-fernet-keys\") pod \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344294 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-dns-svc\") pod \"15970fdb-4e60-4331-bd19-40db152c2303\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344382 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-nb\") pod \"15970fdb-4e60-4331-bd19-40db152c2303\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344402 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkssx\" (UniqueName: \"kubernetes.io/projected/15970fdb-4e60-4331-bd19-40db152c2303-kube-api-access-hkssx\") pod \"15970fdb-4e60-4331-bd19-40db152c2303\" (UID: \"15970fdb-4e60-4331-bd19-40db152c2303\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344430 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-scripts\") pod \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344454 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-config-data\") pod \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344489 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-combined-ca-bundle\") pod \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344530 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-combined-ca-bundle\") pod \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\" (UID: \"abb57ada-d9f9-4162-a8f5-dd3f88a22f50\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.344577 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-config-data\") pod \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\" (UID: \"ef18d029-97e5-4c6b-b93e-0a40f55816a2\") " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.346597 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-logs" (OuterVolumeSpecName: "logs") pod "ef18d029-97e5-4c6b-b93e-0a40f55816a2" (UID: "ef18d029-97e5-4c6b-b93e-0a40f55816a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.351067 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "ef18d029-97e5-4c6b-b93e-0a40f55816a2" (UID: "ef18d029-97e5-4c6b-b93e-0a40f55816a2"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.351514 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-kube-api-access-kc5ft" (OuterVolumeSpecName: "kube-api-access-kc5ft") pod "abb57ada-d9f9-4162-a8f5-dd3f88a22f50" (UID: "abb57ada-d9f9-4162-a8f5-dd3f88a22f50"). InnerVolumeSpecName "kube-api-access-kc5ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.351907 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef18d029-97e5-4c6b-b93e-0a40f55816a2-kube-api-access-wxr56" (OuterVolumeSpecName: "kube-api-access-wxr56") pod "ef18d029-97e5-4c6b-b93e-0a40f55816a2" (UID: "ef18d029-97e5-4c6b-b93e-0a40f55816a2"). InnerVolumeSpecName "kube-api-access-wxr56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.354402 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-scripts" (OuterVolumeSpecName: "scripts") pod "abb57ada-d9f9-4162-a8f5-dd3f88a22f50" (UID: "abb57ada-d9f9-4162-a8f5-dd3f88a22f50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.355453 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "abb57ada-d9f9-4162-a8f5-dd3f88a22f50" (UID: "abb57ada-d9f9-4162-a8f5-dd3f88a22f50"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.357255 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ef18d029-97e5-4c6b-b93e-0a40f55816a2" (UID: "ef18d029-97e5-4c6b-b93e-0a40f55816a2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.359380 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "abb57ada-d9f9-4162-a8f5-dd3f88a22f50" (UID: "abb57ada-d9f9-4162-a8f5-dd3f88a22f50"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.379967 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-scripts" (OuterVolumeSpecName: "scripts") pod "ef18d029-97e5-4c6b-b93e-0a40f55816a2" (UID: "ef18d029-97e5-4c6b-b93e-0a40f55816a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.384622 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-config-data" (OuterVolumeSpecName: "config-data") pod "abb57ada-d9f9-4162-a8f5-dd3f88a22f50" (UID: "abb57ada-d9f9-4162-a8f5-dd3f88a22f50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.386721 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15970fdb-4e60-4331-bd19-40db152c2303-kube-api-access-hkssx" (OuterVolumeSpecName: "kube-api-access-hkssx") pod "15970fdb-4e60-4331-bd19-40db152c2303" (UID: "15970fdb-4e60-4331-bd19-40db152c2303"). InnerVolumeSpecName "kube-api-access-hkssx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.391172 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef18d029-97e5-4c6b-b93e-0a40f55816a2" (UID: "ef18d029-97e5-4c6b-b93e-0a40f55816a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.395212 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abb57ada-d9f9-4162-a8f5-dd3f88a22f50" (UID: "abb57ada-d9f9-4162-a8f5-dd3f88a22f50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.417479 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "15970fdb-4e60-4331-bd19-40db152c2303" (UID: "15970fdb-4e60-4331-bd19-40db152c2303"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.426249 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "15970fdb-4e60-4331-bd19-40db152c2303" (UID: "15970fdb-4e60-4331-bd19-40db152c2303"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.426822 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-config" (OuterVolumeSpecName: "config") pod "15970fdb-4e60-4331-bd19-40db152c2303" (UID: "15970fdb-4e60-4331-bd19-40db152c2303"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.431055 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-config-data" (OuterVolumeSpecName: "config-data") pod "ef18d029-97e5-4c6b-b93e-0a40f55816a2" (UID: "ef18d029-97e5-4c6b-b93e-0a40f55816a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.433575 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "15970fdb-4e60-4331-bd19-40db152c2303" (UID: "15970fdb-4e60-4331-bd19-40db152c2303"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.435438 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ef18d029-97e5-4c6b-b93e-0a40f55816a2" (UID: "ef18d029-97e5-4c6b-b93e-0a40f55816a2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447333 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447385 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447396 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447406 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc5ft\" (UniqueName: \"kubernetes.io/projected/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-kube-api-access-kc5ft\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447416 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxr56\" (UniqueName: \"kubernetes.io/projected/ef18d029-97e5-4c6b-b93e-0a40f55816a2-kube-api-access-wxr56\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447425 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447457 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447466 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447476 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447484 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef18d029-97e5-4c6b-b93e-0a40f55816a2-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447492 4805 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447501 4805 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447509 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447517 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15970fdb-4e60-4331-bd19-40db152c2303-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447524 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkssx\" (UniqueName: \"kubernetes.io/projected/15970fdb-4e60-4331-bd19-40db152c2303-kube-api-access-hkssx\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447532 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447539 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447547 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef18d029-97e5-4c6b-b93e-0a40f55816a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.447555 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb57ada-d9f9-4162-a8f5-dd3f88a22f50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.477665 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.549678 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.894732 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" event={"ID":"15970fdb-4e60-4331-bd19-40db152c2303","Type":"ContainerDied","Data":"2429800ddd1c05e9b28bf98cfe741ef2272cec1c76ae35135631a81fc74b7f7b"} Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.894797 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.897523 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ef18d029-97e5-4c6b-b93e-0a40f55816a2","Type":"ContainerDied","Data":"3d9a85980d71759f88a5ed4cb21763aea8c50c82bc1b3c452cdd0a35cfbe3cf5"} Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.897578 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.909044 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fj49w" event={"ID":"abb57ada-d9f9-4162-a8f5-dd3f88a22f50","Type":"ContainerDied","Data":"bb14b64653965cdb12040c9c92f15b133dca4eb5cdc076c177bea0a664a8e0bd"} Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.909094 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb14b64653965cdb12040c9c92f15b133dca4eb5cdc076c177bea0a664a8e0bd" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.909176 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fj49w" Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.938996 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-4lt76"] Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.952027 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-4lt76"] Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.972936 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:45:58 crc kubenswrapper[4805]: I1128 15:45:58.987539 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.000090 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:45:59 crc kubenswrapper[4805]: E1128 15:45:59.000582 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerName="glance-log" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.000608 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerName="glance-log" Nov 28 15:45:59 crc kubenswrapper[4805]: E1128 15:45:59.000623 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb57ada-d9f9-4162-a8f5-dd3f88a22f50" containerName="keystone-bootstrap" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.000633 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb57ada-d9f9-4162-a8f5-dd3f88a22f50" containerName="keystone-bootstrap" Nov 28 15:45:59 crc kubenswrapper[4805]: E1128 15:45:59.000667 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="dnsmasq-dns" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.000675 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="dnsmasq-dns" Nov 28 15:45:59 crc kubenswrapper[4805]: E1128 15:45:59.000697 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerName="glance-httpd" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.000705 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerName="glance-httpd" Nov 28 15:45:59 crc kubenswrapper[4805]: E1128 15:45:59.000719 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4322e7fd-5397-4d32-a24c-1c3b299171a5" containerName="init" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.000727 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4322e7fd-5397-4d32-a24c-1c3b299171a5" containerName="init" Nov 28 15:45:59 crc kubenswrapper[4805]: E1128 15:45:59.000737 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="init" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.000744 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="init" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.000954 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb57ada-d9f9-4162-a8f5-dd3f88a22f50" containerName="keystone-bootstrap" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.000978 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerName="glance-log" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.001000 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4322e7fd-5397-4d32-a24c-1c3b299171a5" containerName="init" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.001010 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="dnsmasq-dns" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.001024 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" containerName="glance-httpd" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.009744 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.009880 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.011957 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.013997 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.158232 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.158302 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.158333 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.158394 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.158439 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-logs\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.158464 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.158518 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.158655 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75lzn\" (UniqueName: \"kubernetes.io/projected/b9d91b74-236c-4989-93ca-2379328e7119-kube-api-access-75lzn\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.218025 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15970fdb-4e60-4331-bd19-40db152c2303" path="/var/lib/kubelet/pods/15970fdb-4e60-4331-bd19-40db152c2303/volumes" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.218698 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef18d029-97e5-4c6b-b93e-0a40f55816a2" path="/var/lib/kubelet/pods/ef18d029-97e5-4c6b-b93e-0a40f55816a2/volumes" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.259943 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.260003 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.260054 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.260098 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-logs\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.260124 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.260180 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.260241 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75lzn\" (UniqueName: \"kubernetes.io/projected/b9d91b74-236c-4989-93ca-2379328e7119-kube-api-access-75lzn\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.260298 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.260744 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.260770 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-logs\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.262475 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.263816 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.264701 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.265311 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.265809 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.298250 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.300731 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75lzn\" (UniqueName: \"kubernetes.io/projected/b9d91b74-236c-4989-93ca-2379328e7119-kube-api-access-75lzn\") pod \"glance-default-internal-api-0\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.336491 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.418419 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fj49w"] Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.431555 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fj49w"] Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.474189 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-vqdvs"] Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.475596 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.481911 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.482073 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.482296 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.482600 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.482835 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fzt55" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.488842 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vqdvs"] Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.571988 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-scripts\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.572035 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-credential-keys\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.572081 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-fernet-keys\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.572113 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx62j\" (UniqueName: \"kubernetes.io/projected/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-kube-api-access-nx62j\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.572182 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-config-data\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.572200 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-combined-ca-bundle\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.675572 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-scripts\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.675610 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-credential-keys\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.675657 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-fernet-keys\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.675687 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx62j\" (UniqueName: \"kubernetes.io/projected/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-kube-api-access-nx62j\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.675726 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-config-data\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.675769 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-combined-ca-bundle\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.680108 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-combined-ca-bundle\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.680326 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-scripts\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.680326 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-credential-keys\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.688268 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-config-data\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.688749 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-fernet-keys\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.690693 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx62j\" (UniqueName: \"kubernetes.io/projected/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-kube-api-access-nx62j\") pod \"keystone-bootstrap-vqdvs\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.762550 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f6d79597f-4lt76" podUID="15970fdb-4e60-4331-bd19-40db152c2303" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Nov 28 15:45:59 crc kubenswrapper[4805]: I1128 15:45:59.831101 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.111451 4805 scope.go:117] "RemoveContainer" containerID="8f6f0afc837b022cd04e980a4e338bee80b78148bdf3b45649b526488e1ea13c" Nov 28 15:46:00 crc kubenswrapper[4805]: E1128 15:46:00.151569 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b5266c9a26766fce2b92f95dff52d362a760f7baf1474cdcb33bd68570e096c0" Nov 28 15:46:00 crc kubenswrapper[4805]: E1128 15:46:00.151727 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b5266c9a26766fce2b92f95dff52d362a760f7baf1474cdcb33bd68570e096c0,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9bvd5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-zwp6n_openstack(b2f218fa-9d63-46f4-ba0d-3b9166d23672): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 15:46:00 crc kubenswrapper[4805]: E1128 15:46:00.153075 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-zwp6n" podUID="b2f218fa-9d63-46f4-ba0d-3b9166d23672" Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.187468 4805 scope.go:117] "RemoveContainer" containerID="96ccfd45f61b250ae5d9a05706495e9330c8aa06717bc09526eff352434d44bf" Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.250600 4805 scope.go:117] "RemoveContainer" containerID="05de3845991eaa359e6baa626dc02171b7f292526b958ada1028d01c0e1669d6" Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.371874 4805 scope.go:117] "RemoveContainer" containerID="8de05ca7dc13db2d90ef6ba4f2b7f9de29b8b13fb05bcb06f8eb51f955edf2a3" Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.492143 4805 scope.go:117] "RemoveContainer" containerID="3bdc559c6aa17f08eca4f8a4b86e51ca90fd3eb7ad664f71ba07dd24f99a07d6" Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.543589 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vqdvs"] Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.553603 4805 scope.go:117] "RemoveContainer" containerID="5a22dde5461cfab341af8734383933aea39409c34b8df382b72cc7579b8593fe" Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.690821 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.888366 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:46:00 crc kubenswrapper[4805]: W1128 15:46:00.897780 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9d91b74_236c_4989_93ca_2379328e7119.slice/crio-e97cb1db4f9760c1fe5620f60e7cc8cc379267d2ac3cea0acf084549a08d85be WatchSource:0}: Error finding container e97cb1db4f9760c1fe5620f60e7cc8cc379267d2ac3cea0acf084549a08d85be: Status 404 returned error can't find the container with id e97cb1db4f9760c1fe5620f60e7cc8cc379267d2ac3cea0acf084549a08d85be Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.944789 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vqdvs" event={"ID":"67c9ab51-9cfd-4e5f-9df6-1f226ab94173","Type":"ContainerStarted","Data":"8bdacf735cc55e4eb9afb66b2e37c11f6d54e6bb6bec04837719d6ce7e4e5af8"} Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.944830 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vqdvs" event={"ID":"67c9ab51-9cfd-4e5f-9df6-1f226ab94173","Type":"ContainerStarted","Data":"3b2b0c4d9c13c23be42d1677ec35e7954e0c7290988566cecd880b7533f02c03"} Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.949271 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17e2c4d9-4960-4034-be79-5dffbab66c5b","Type":"ContainerStarted","Data":"a9d59ed5f828699fd3b23a41a47fc4b3d4c4e17931ca38678b23a5d7a1e55fda"} Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.950934 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hbmtq" event={"ID":"37bb6d65-078d-4d19-bc10-dc50ea71edae","Type":"ContainerStarted","Data":"14da7f9c8fb9e23ab79dd5e17973e13aa4fe285d149b8b95925f3a3f92426818"} Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.964320 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-kfvps" event={"ID":"e63a3a44-329f-4d06-af10-a8ac7c72dcc4","Type":"ContainerStarted","Data":"6260afcad67d63a4d5d64b9acce81ba750c382607d4f5264abef41c12d820237"} Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.967570 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-vqdvs" podStartSLOduration=1.967550652 podStartE2EDuration="1.967550652s" podCreationTimestamp="2025-11-28 15:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:00.963499321 +0000 UTC m=+1188.013290632" watchObservedRunningTime="2025-11-28 15:46:00.967550652 +0000 UTC m=+1188.017341963" Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.988726 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b9d91b74-236c-4989-93ca-2379328e7119","Type":"ContainerStarted","Data":"e97cb1db4f9760c1fe5620f60e7cc8cc379267d2ac3cea0acf084549a08d85be"} Nov 28 15:46:00 crc kubenswrapper[4805]: I1128 15:46:00.994501 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"122b157a-374d-437e-97bb-66fa032893ff","Type":"ContainerStarted","Data":"212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66"} Nov 28 15:46:00 crc kubenswrapper[4805]: E1128 15:46:00.996225 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b5266c9a26766fce2b92f95dff52d362a760f7baf1474cdcb33bd68570e096c0\\\"\"" pod="openstack/cinder-db-sync-zwp6n" podUID="b2f218fa-9d63-46f4-ba0d-3b9166d23672" Nov 28 15:46:01 crc kubenswrapper[4805]: I1128 15:46:01.014497 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-hbmtq" podStartSLOduration=2.329285957 podStartE2EDuration="32.014479949s" podCreationTimestamp="2025-11-28 15:45:29 +0000 UTC" firstStartedPulling="2025-11-28 15:45:30.460983128 +0000 UTC m=+1157.510774449" lastFinishedPulling="2025-11-28 15:46:00.14617713 +0000 UTC m=+1187.195968441" observedRunningTime="2025-11-28 15:46:00.98771025 +0000 UTC m=+1188.037501571" watchObservedRunningTime="2025-11-28 15:46:01.014479949 +0000 UTC m=+1188.064271260" Nov 28 15:46:01 crc kubenswrapper[4805]: I1128 15:46:01.017005 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-kfvps" podStartSLOduration=2.459968604 podStartE2EDuration="32.016994677s" podCreationTimestamp="2025-11-28 15:45:29 +0000 UTC" firstStartedPulling="2025-11-28 15:45:30.58562375 +0000 UTC m=+1157.635415061" lastFinishedPulling="2025-11-28 15:46:00.142649823 +0000 UTC m=+1187.192441134" observedRunningTime="2025-11-28 15:46:01.009118123 +0000 UTC m=+1188.058909454" watchObservedRunningTime="2025-11-28 15:46:01.016994677 +0000 UTC m=+1188.066785988" Nov 28 15:46:01 crc kubenswrapper[4805]: I1128 15:46:01.222948 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abb57ada-d9f9-4162-a8f5-dd3f88a22f50" path="/var/lib/kubelet/pods/abb57ada-d9f9-4162-a8f5-dd3f88a22f50/volumes" Nov 28 15:46:02 crc kubenswrapper[4805]: I1128 15:46:02.010462 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b9d91b74-236c-4989-93ca-2379328e7119","Type":"ContainerStarted","Data":"c655e1242a0c7987b02a1c84687637804a2b82ea438583fed89c15a2e0f3dc56"} Nov 28 15:46:02 crc kubenswrapper[4805]: I1128 15:46:02.012726 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17e2c4d9-4960-4034-be79-5dffbab66c5b","Type":"ContainerStarted","Data":"3132a6f16b1d3aee6579746b5437f124996d0c2d9789b01ebda38770a8012732"} Nov 28 15:46:03 crc kubenswrapper[4805]: I1128 15:46:03.022720 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b9d91b74-236c-4989-93ca-2379328e7119","Type":"ContainerStarted","Data":"1af8fb2da9461341a610764bda29e02ea05cb8bafd6f608ea0e1266702bb14e5"} Nov 28 15:46:03 crc kubenswrapper[4805]: I1128 15:46:03.024738 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17e2c4d9-4960-4034-be79-5dffbab66c5b","Type":"ContainerStarted","Data":"b050f44a1b2eee2ec7d7f2eaa9e0af5107a6d79deecade5b296d226400c82649"} Nov 28 15:46:03 crc kubenswrapper[4805]: I1128 15:46:03.026961 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"122b157a-374d-437e-97bb-66fa032893ff","Type":"ContainerStarted","Data":"4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd"} Nov 28 15:46:03 crc kubenswrapper[4805]: I1128 15:46:03.054754 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.054735219 podStartE2EDuration="5.054735219s" podCreationTimestamp="2025-11-28 15:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:03.051139612 +0000 UTC m=+1190.100930923" watchObservedRunningTime="2025-11-28 15:46:03.054735219 +0000 UTC m=+1190.104526530" Nov 28 15:46:03 crc kubenswrapper[4805]: I1128 15:46:03.087927 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=23.087908772 podStartE2EDuration="23.087908772s" podCreationTimestamp="2025-11-28 15:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:03.085674231 +0000 UTC m=+1190.135465542" watchObservedRunningTime="2025-11-28 15:46:03.087908772 +0000 UTC m=+1190.137700083" Nov 28 15:46:05 crc kubenswrapper[4805]: I1128 15:46:05.046431 4805 generic.go:334] "Generic (PLEG): container finished" podID="67c9ab51-9cfd-4e5f-9df6-1f226ab94173" containerID="8bdacf735cc55e4eb9afb66b2e37c11f6d54e6bb6bec04837719d6ce7e4e5af8" exitCode=0 Nov 28 15:46:05 crc kubenswrapper[4805]: I1128 15:46:05.046539 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vqdvs" event={"ID":"67c9ab51-9cfd-4e5f-9df6-1f226ab94173","Type":"ContainerDied","Data":"8bdacf735cc55e4eb9afb66b2e37c11f6d54e6bb6bec04837719d6ce7e4e5af8"} Nov 28 15:46:06 crc kubenswrapper[4805]: I1128 15:46:06.058625 4805 generic.go:334] "Generic (PLEG): container finished" podID="37bb6d65-078d-4d19-bc10-dc50ea71edae" containerID="14da7f9c8fb9e23ab79dd5e17973e13aa4fe285d149b8b95925f3a3f92426818" exitCode=0 Nov 28 15:46:06 crc kubenswrapper[4805]: I1128 15:46:06.058811 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hbmtq" event={"ID":"37bb6d65-078d-4d19-bc10-dc50ea71edae","Type":"ContainerDied","Data":"14da7f9c8fb9e23ab79dd5e17973e13aa4fe285d149b8b95925f3a3f92426818"} Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.239748 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.327082 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-config-data\") pod \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.327179 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-combined-ca-bundle\") pod \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.327268 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-credential-keys\") pod \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.327490 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-scripts\") pod \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.327534 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-fernet-keys\") pod \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.327570 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx62j\" (UniqueName: \"kubernetes.io/projected/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-kube-api-access-nx62j\") pod \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\" (UID: \"67c9ab51-9cfd-4e5f-9df6-1f226ab94173\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.341547 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "67c9ab51-9cfd-4e5f-9df6-1f226ab94173" (UID: "67c9ab51-9cfd-4e5f-9df6-1f226ab94173"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.347597 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-scripts" (OuterVolumeSpecName: "scripts") pod "67c9ab51-9cfd-4e5f-9df6-1f226ab94173" (UID: "67c9ab51-9cfd-4e5f-9df6-1f226ab94173"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.357194 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-kube-api-access-nx62j" (OuterVolumeSpecName: "kube-api-access-nx62j") pod "67c9ab51-9cfd-4e5f-9df6-1f226ab94173" (UID: "67c9ab51-9cfd-4e5f-9df6-1f226ab94173"). InnerVolumeSpecName "kube-api-access-nx62j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.378541 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "67c9ab51-9cfd-4e5f-9df6-1f226ab94173" (UID: "67c9ab51-9cfd-4e5f-9df6-1f226ab94173"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.429862 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.429903 4805 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.429918 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx62j\" (UniqueName: \"kubernetes.io/projected/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-kube-api-access-nx62j\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.429930 4805 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.438585 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-config-data" (OuterVolumeSpecName: "config-data") pod "67c9ab51-9cfd-4e5f-9df6-1f226ab94173" (UID: "67c9ab51-9cfd-4e5f-9df6-1f226ab94173"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.449059 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67c9ab51-9cfd-4e5f-9df6-1f226ab94173" (UID: "67c9ab51-9cfd-4e5f-9df6-1f226ab94173"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.520980 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hbmtq" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.531724 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.531751 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67c9ab51-9cfd-4e5f-9df6-1f226ab94173-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.632589 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-scripts\") pod \"37bb6d65-078d-4d19-bc10-dc50ea71edae\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.632716 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-config-data\") pod \"37bb6d65-078d-4d19-bc10-dc50ea71edae\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.632746 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37bb6d65-078d-4d19-bc10-dc50ea71edae-logs\") pod \"37bb6d65-078d-4d19-bc10-dc50ea71edae\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.633143 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37bb6d65-078d-4d19-bc10-dc50ea71edae-logs" (OuterVolumeSpecName: "logs") pod "37bb6d65-078d-4d19-bc10-dc50ea71edae" (UID: "37bb6d65-078d-4d19-bc10-dc50ea71edae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.633225 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sffpv\" (UniqueName: \"kubernetes.io/projected/37bb6d65-078d-4d19-bc10-dc50ea71edae-kube-api-access-sffpv\") pod \"37bb6d65-078d-4d19-bc10-dc50ea71edae\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.633307 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-combined-ca-bundle\") pod \"37bb6d65-078d-4d19-bc10-dc50ea71edae\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.633774 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37bb6d65-078d-4d19-bc10-dc50ea71edae-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.638561 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37bb6d65-078d-4d19-bc10-dc50ea71edae-kube-api-access-sffpv" (OuterVolumeSpecName: "kube-api-access-sffpv") pod "37bb6d65-078d-4d19-bc10-dc50ea71edae" (UID: "37bb6d65-078d-4d19-bc10-dc50ea71edae"). InnerVolumeSpecName "kube-api-access-sffpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.641189 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-scripts" (OuterVolumeSpecName: "scripts") pod "37bb6d65-078d-4d19-bc10-dc50ea71edae" (UID: "37bb6d65-078d-4d19-bc10-dc50ea71edae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: E1128 15:46:07.673794 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-combined-ca-bundle podName:37bb6d65-078d-4d19-bc10-dc50ea71edae nodeName:}" failed. No retries permitted until 2025-11-28 15:46:08.173765175 +0000 UTC m=+1195.223556476 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-combined-ca-bundle") pod "37bb6d65-078d-4d19-bc10-dc50ea71edae" (UID: "37bb6d65-078d-4d19-bc10-dc50ea71edae") : error deleting /var/lib/kubelet/pods/37bb6d65-078d-4d19-bc10-dc50ea71edae/volume-subpaths: remove /var/lib/kubelet/pods/37bb6d65-078d-4d19-bc10-dc50ea71edae/volume-subpaths: no such file or directory Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.677404 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-config-data" (OuterVolumeSpecName: "config-data") pod "37bb6d65-078d-4d19-bc10-dc50ea71edae" (UID: "37bb6d65-078d-4d19-bc10-dc50ea71edae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.735878 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.735912 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:07 crc kubenswrapper[4805]: I1128 15:46:07.735923 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sffpv\" (UniqueName: \"kubernetes.io/projected/37bb6d65-078d-4d19-bc10-dc50ea71edae-kube-api-access-sffpv\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.109521 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hbmtq" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.109511 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hbmtq" event={"ID":"37bb6d65-078d-4d19-bc10-dc50ea71edae","Type":"ContainerDied","Data":"fcfc234ef249807518798b35eac6dcf6ff04a0eead585c05895e89101d2c4432"} Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.109895 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcfc234ef249807518798b35eac6dcf6ff04a0eead585c05895e89101d2c4432" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.111541 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vqdvs" event={"ID":"67c9ab51-9cfd-4e5f-9df6-1f226ab94173","Type":"ContainerDied","Data":"3b2b0c4d9c13c23be42d1677ec35e7954e0c7290988566cecd880b7533f02c03"} Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.111569 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b2b0c4d9c13c23be42d1677ec35e7954e0c7290988566cecd880b7533f02c03" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.111625 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vqdvs" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.118066 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"122b157a-374d-437e-97bb-66fa032893ff","Type":"ContainerStarted","Data":"a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43"} Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.206656 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-55d54b965d-bfqn6"] Nov 28 15:46:08 crc kubenswrapper[4805]: E1128 15:46:08.207466 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37bb6d65-078d-4d19-bc10-dc50ea71edae" containerName="placement-db-sync" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.207489 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="37bb6d65-078d-4d19-bc10-dc50ea71edae" containerName="placement-db-sync" Nov 28 15:46:08 crc kubenswrapper[4805]: E1128 15:46:08.207520 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c9ab51-9cfd-4e5f-9df6-1f226ab94173" containerName="keystone-bootstrap" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.207529 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c9ab51-9cfd-4e5f-9df6-1f226ab94173" containerName="keystone-bootstrap" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.207726 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="37bb6d65-078d-4d19-bc10-dc50ea71edae" containerName="placement-db-sync" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.207758 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="67c9ab51-9cfd-4e5f-9df6-1f226ab94173" containerName="keystone-bootstrap" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.220874 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-55d54b965d-bfqn6"] Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.221004 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.223840 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.224137 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.249940 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-combined-ca-bundle\") pod \"37bb6d65-078d-4d19-bc10-dc50ea71edae\" (UID: \"37bb6d65-078d-4d19-bc10-dc50ea71edae\") " Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.255431 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37bb6d65-078d-4d19-bc10-dc50ea71edae" (UID: "37bb6d65-078d-4d19-bc10-dc50ea71edae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.352240 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-scripts\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.352329 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-config-data\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.352368 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-combined-ca-bundle\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.352465 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-public-tls-certs\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.352494 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-logs\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.352549 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hldcl\" (UniqueName: \"kubernetes.io/projected/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-kube-api-access-hldcl\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.352626 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-internal-tls-certs\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.352686 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb6d65-078d-4d19-bc10-dc50ea71edae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.372602 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5f69cf9fcb-rvfkd"] Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.374064 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.377310 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.377780 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.377977 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.378028 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.378307 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.378382 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fzt55" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.383217 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5f69cf9fcb-rvfkd"] Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.453644 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-combined-ca-bundle\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.453681 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-logs\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.453731 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hldcl\" (UniqueName: \"kubernetes.io/projected/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-kube-api-access-hldcl\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454105 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-logs\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454507 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-internal-tls-certs\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454541 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-scripts\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454710 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-config-data\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454758 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-combined-ca-bundle\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454823 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-scripts\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454855 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-fernet-keys\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454919 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-credential-keys\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454962 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-internal-tls-certs\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.454997 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xpv2\" (UniqueName: \"kubernetes.io/projected/614a68ec-3129-413f-abb1-40a73ad9137e-kube-api-access-2xpv2\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.455023 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-config-data\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.455094 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-public-tls-certs\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.455127 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-public-tls-certs\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.458647 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-internal-tls-certs\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.458837 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-config-data\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.459155 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-scripts\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.460430 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-combined-ca-bundle\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.471312 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-public-tls-certs\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.471554 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hldcl\" (UniqueName: \"kubernetes.io/projected/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-kube-api-access-hldcl\") pod \"placement-55d54b965d-bfqn6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.537271 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.556818 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-scripts\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.556888 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-fernet-keys\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.556925 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-credential-keys\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.556952 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-internal-tls-certs\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.556984 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xpv2\" (UniqueName: \"kubernetes.io/projected/614a68ec-3129-413f-abb1-40a73ad9137e-kube-api-access-2xpv2\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.557007 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-config-data\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.557042 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-public-tls-certs\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.557064 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-combined-ca-bundle\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.560754 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-combined-ca-bundle\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.563077 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-fernet-keys\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.563143 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-internal-tls-certs\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.563465 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-config-data\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.565492 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-credential-keys\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.566903 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-scripts\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.575447 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-public-tls-certs\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.583192 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xpv2\" (UniqueName: \"kubernetes.io/projected/614a68ec-3129-413f-abb1-40a73ad9137e-kube-api-access-2xpv2\") pod \"keystone-5f69cf9fcb-rvfkd\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:08 crc kubenswrapper[4805]: I1128 15:46:08.698677 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:09 crc kubenswrapper[4805]: I1128 15:46:09.040312 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-55d54b965d-bfqn6"] Nov 28 15:46:09 crc kubenswrapper[4805]: I1128 15:46:09.127338 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55d54b965d-bfqn6" event={"ID":"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6","Type":"ContainerStarted","Data":"1de6d6193c41094be9eb487181477d290298455ffdb3ad015fa40fffe8dd265f"} Nov 28 15:46:09 crc kubenswrapper[4805]: I1128 15:46:09.129974 4805 generic.go:334] "Generic (PLEG): container finished" podID="e63a3a44-329f-4d06-af10-a8ac7c72dcc4" containerID="6260afcad67d63a4d5d64b9acce81ba750c382607d4f5264abef41c12d820237" exitCode=0 Nov 28 15:46:09 crc kubenswrapper[4805]: I1128 15:46:09.130006 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-kfvps" event={"ID":"e63a3a44-329f-4d06-af10-a8ac7c72dcc4","Type":"ContainerDied","Data":"6260afcad67d63a4d5d64b9acce81ba750c382607d4f5264abef41c12d820237"} Nov 28 15:46:09 crc kubenswrapper[4805]: W1128 15:46:09.184075 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod614a68ec_3129_413f_abb1_40a73ad9137e.slice/crio-5e0db73f65439bf234434ee8cbf8f01e05605fc1ebe5e373a0f6994435fb3c43 WatchSource:0}: Error finding container 5e0db73f65439bf234434ee8cbf8f01e05605fc1ebe5e373a0f6994435fb3c43: Status 404 returned error can't find the container with id 5e0db73f65439bf234434ee8cbf8f01e05605fc1ebe5e373a0f6994435fb3c43 Nov 28 15:46:09 crc kubenswrapper[4805]: I1128 15:46:09.189184 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5f69cf9fcb-rvfkd"] Nov 28 15:46:09 crc kubenswrapper[4805]: I1128 15:46:09.336989 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 15:46:09 crc kubenswrapper[4805]: I1128 15:46:09.337042 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 15:46:09 crc kubenswrapper[4805]: I1128 15:46:09.370324 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 15:46:09 crc kubenswrapper[4805]: I1128 15:46:09.381286 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.169252 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55d54b965d-bfqn6" event={"ID":"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6","Type":"ContainerStarted","Data":"fe684bfa9591b55be6b83497c5f2059493fe4abdfa52867aaec588e8310d8893"} Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.169899 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55d54b965d-bfqn6" event={"ID":"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6","Type":"ContainerStarted","Data":"3662b2162435ede3ed240fafeef413deeef6d00aecf000c0f5887352f3c5a777"} Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.170590 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.170638 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.189501 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f69cf9fcb-rvfkd" event={"ID":"614a68ec-3129-413f-abb1-40a73ad9137e","Type":"ContainerStarted","Data":"28e9827ec5a663c00f28c55d8c943235e37131e6dc45f5d050381a2a732dd392"} Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.189545 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.189559 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f69cf9fcb-rvfkd" event={"ID":"614a68ec-3129-413f-abb1-40a73ad9137e","Type":"ContainerStarted","Data":"5e0db73f65439bf234434ee8cbf8f01e05605fc1ebe5e373a0f6994435fb3c43"} Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.189683 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.190101 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.199403 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-55d54b965d-bfqn6" podStartSLOduration=2.199378003 podStartE2EDuration="2.199378003s" podCreationTimestamp="2025-11-28 15:46:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:10.19778263 +0000 UTC m=+1197.247573971" watchObservedRunningTime="2025-11-28 15:46:10.199378003 +0000 UTC m=+1197.249169314" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.252004 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5f69cf9fcb-rvfkd" podStartSLOduration=2.251985465 podStartE2EDuration="2.251985465s" podCreationTimestamp="2025-11-28 15:46:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:10.212736117 +0000 UTC m=+1197.262527428" watchObservedRunningTime="2025-11-28 15:46:10.251985465 +0000 UTC m=+1197.301776776" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.657829 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-kfvps" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.716464 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-db-sync-config-data\") pod \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.716513 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4c8q\" (UniqueName: \"kubernetes.io/projected/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-kube-api-access-f4c8q\") pod \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.716622 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-combined-ca-bundle\") pod \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\" (UID: \"e63a3a44-329f-4d06-af10-a8ac7c72dcc4\") " Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.725461 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e63a3a44-329f-4d06-af10-a8ac7c72dcc4" (UID: "e63a3a44-329f-4d06-af10-a8ac7c72dcc4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.734814 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-kube-api-access-f4c8q" (OuterVolumeSpecName: "kube-api-access-f4c8q") pod "e63a3a44-329f-4d06-af10-a8ac7c72dcc4" (UID: "e63a3a44-329f-4d06-af10-a8ac7c72dcc4"). InnerVolumeSpecName "kube-api-access-f4c8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.765502 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e63a3a44-329f-4d06-af10-a8ac7c72dcc4" (UID: "e63a3a44-329f-4d06-af10-a8ac7c72dcc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.818287 4805 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.818328 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4c8q\" (UniqueName: \"kubernetes.io/projected/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-kube-api-access-f4c8q\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:10 crc kubenswrapper[4805]: I1128 15:46:10.818341 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63a3a44-329f-4d06-af10-a8ac7c72dcc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.126018 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.126409 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.126424 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.126435 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.171587 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.187516 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.195349 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-kfvps" event={"ID":"e63a3a44-329f-4d06-af10-a8ac7c72dcc4","Type":"ContainerDied","Data":"a14d597375e8964a17ca4d513211ec5b60b710e3572322748d5b95c1f34d5673"} Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.195391 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a14d597375e8964a17ca4d513211ec5b60b710e3572322748d5b95c1f34d5673" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.195534 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-kfvps" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.409106 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4"] Nov 28 15:46:11 crc kubenswrapper[4805]: E1128 15:46:11.409467 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e63a3a44-329f-4d06-af10-a8ac7c72dcc4" containerName="barbican-db-sync" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.409480 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e63a3a44-329f-4d06-af10-a8ac7c72dcc4" containerName="barbican-db-sync" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.409818 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e63a3a44-329f-4d06-af10-a8ac7c72dcc4" containerName="barbican-db-sync" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.413263 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.423155 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4"] Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.423309 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.423679 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gzh2w" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.429718 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.474523 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-97ccb4d45-ddn2r"] Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.476424 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.481426 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529496 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529564 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data-custom\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529604 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529696 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-combined-ca-bundle\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529742 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abfd499c-e83b-4616-a80c-29a7e8e750e2-logs\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529776 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st67h\" (UniqueName: \"kubernetes.io/projected/a363e2d1-0a91-4578-b9e4-b736b7931b03-kube-api-access-st67h\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529800 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ftr4\" (UniqueName: \"kubernetes.io/projected/abfd499c-e83b-4616-a80c-29a7e8e750e2-kube-api-access-2ftr4\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529826 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a363e2d1-0a91-4578-b9e4-b736b7931b03-logs\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529888 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data-custom\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.529928 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-combined-ca-bundle\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.541550 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-97ccb4d45-ddn2r"] Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.581430 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bcffb858c-v9wlq"] Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.589283 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.608153 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bcffb858c-v9wlq"] Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.631769 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-svc\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.632027 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data-custom\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.632128 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-combined-ca-bundle\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.632213 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-nb\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.632305 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.632437 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data-custom\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.632532 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.632640 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l47rq\" (UniqueName: \"kubernetes.io/projected/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-kube-api-access-l47rq\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.632728 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-sb\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.633527 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-combined-ca-bundle\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.633626 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abfd499c-e83b-4616-a80c-29a7e8e750e2-logs\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.633663 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-config\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.633697 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st67h\" (UniqueName: \"kubernetes.io/projected/a363e2d1-0a91-4578-b9e4-b736b7931b03-kube-api-access-st67h\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.633730 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ftr4\" (UniqueName: \"kubernetes.io/projected/abfd499c-e83b-4616-a80c-29a7e8e750e2-kube-api-access-2ftr4\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.633751 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-swift-storage-0\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.633780 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a363e2d1-0a91-4578-b9e4-b736b7931b03-logs\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.634252 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a363e2d1-0a91-4578-b9e4-b736b7931b03-logs\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.643234 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abfd499c-e83b-4616-a80c-29a7e8e750e2-logs\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.643530 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-combined-ca-bundle\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.646003 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data-custom\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.648010 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.648336 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-combined-ca-bundle\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.648949 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data-custom\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.653849 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6d99f7798d-mk9wl"] Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.655242 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.678099 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.683470 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ftr4\" (UniqueName: \"kubernetes.io/projected/abfd499c-e83b-4616-a80c-29a7e8e750e2-kube-api-access-2ftr4\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.687183 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d99f7798d-mk9wl"] Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.693039 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data\") pod \"barbican-worker-97ccb4d45-ddn2r\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.693872 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st67h\" (UniqueName: \"kubernetes.io/projected/a363e2d1-0a91-4578-b9e4-b736b7931b03-kube-api-access-st67h\") pod \"barbican-keystone-listener-5c6d5f89c4-q6gr4\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.740609 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l47rq\" (UniqueName: \"kubernetes.io/projected/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-kube-api-access-l47rq\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.740729 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-sb\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.740783 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data-custom\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.740897 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-config\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.740933 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jcjj\" (UniqueName: \"kubernetes.io/projected/51176eca-5fa0-411d-976f-4095550240cd-kube-api-access-6jcjj\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.740969 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-swift-storage-0\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.741011 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51176eca-5fa0-411d-976f-4095550240cd-logs\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.741038 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-svc\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.741119 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-combined-ca-bundle\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.741161 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.741203 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-nb\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.742720 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-svc\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.742931 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-nb\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.743147 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-sb\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.743318 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.748026 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-config\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.753693 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-swift-storage-0\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.777936 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l47rq\" (UniqueName: \"kubernetes.io/projected/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-kube-api-access-l47rq\") pod \"dnsmasq-dns-bcffb858c-v9wlq\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.805915 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.843508 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data-custom\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.843683 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jcjj\" (UniqueName: \"kubernetes.io/projected/51176eca-5fa0-411d-976f-4095550240cd-kube-api-access-6jcjj\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.843804 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51176eca-5fa0-411d-976f-4095550240cd-logs\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.843894 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-combined-ca-bundle\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.843927 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.849336 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51176eca-5fa0-411d-976f-4095550240cd-logs\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.853439 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-combined-ca-bundle\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.853972 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.863097 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data-custom\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.878138 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jcjj\" (UniqueName: \"kubernetes.io/projected/51176eca-5fa0-411d-976f-4095550240cd-kube-api-access-6jcjj\") pod \"barbican-api-6d99f7798d-mk9wl\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:11 crc kubenswrapper[4805]: I1128 15:46:11.922050 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:12 crc kubenswrapper[4805]: I1128 15:46:12.150939 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:12 crc kubenswrapper[4805]: I1128 15:46:12.227438 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:46:12 crc kubenswrapper[4805]: I1128 15:46:12.227473 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:46:12 crc kubenswrapper[4805]: I1128 15:46:12.307558 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-97ccb4d45-ddn2r"] Nov 28 15:46:12 crc kubenswrapper[4805]: W1128 15:46:12.358085 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabfd499c_e83b_4616_a80c_29a7e8e750e2.slice/crio-0f69434a9db9bceb12c66989fba99a7fce7624aaa43608837cd7f1d51a750edb WatchSource:0}: Error finding container 0f69434a9db9bceb12c66989fba99a7fce7624aaa43608837cd7f1d51a750edb: Status 404 returned error can't find the container with id 0f69434a9db9bceb12c66989fba99a7fce7624aaa43608837cd7f1d51a750edb Nov 28 15:46:12 crc kubenswrapper[4805]: I1128 15:46:12.502992 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4"] Nov 28 15:46:12 crc kubenswrapper[4805]: I1128 15:46:12.658283 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bcffb858c-v9wlq"] Nov 28 15:46:12 crc kubenswrapper[4805]: I1128 15:46:12.785073 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d99f7798d-mk9wl"] Nov 28 15:46:13 crc kubenswrapper[4805]: I1128 15:46:13.248725 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 15:46:13 crc kubenswrapper[4805]: I1128 15:46:13.249773 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 15:46:13 crc kubenswrapper[4805]: I1128 15:46:13.256205 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d99f7798d-mk9wl" event={"ID":"51176eca-5fa0-411d-976f-4095550240cd","Type":"ContainerStarted","Data":"2dd22d1165298bd8fae99be54a7da6131f198cf89ebec17a42db3f5a79bec75c"} Nov 28 15:46:13 crc kubenswrapper[4805]: I1128 15:46:13.256249 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d99f7798d-mk9wl" event={"ID":"51176eca-5fa0-411d-976f-4095550240cd","Type":"ContainerStarted","Data":"6d7348aaafe5568f2fc6168ba682b8272118d91304a13f8b6dd42e7276d0f5e4"} Nov 28 15:46:13 crc kubenswrapper[4805]: I1128 15:46:13.284741 4805 generic.go:334] "Generic (PLEG): container finished" podID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerID="18e7dc78edf79d28cdfed7f21eb157b3e47abb87593b65dc74d696841dc2a324" exitCode=0 Nov 28 15:46:13 crc kubenswrapper[4805]: I1128 15:46:13.284821 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" event={"ID":"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950","Type":"ContainerDied","Data":"18e7dc78edf79d28cdfed7f21eb157b3e47abb87593b65dc74d696841dc2a324"} Nov 28 15:46:13 crc kubenswrapper[4805]: I1128 15:46:13.284847 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" event={"ID":"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950","Type":"ContainerStarted","Data":"4f513c6ed66741f0cd62c728193c47e57852b094a4b69d6b76193ac63582f460"} Nov 28 15:46:13 crc kubenswrapper[4805]: I1128 15:46:13.315550 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-97ccb4d45-ddn2r" event={"ID":"abfd499c-e83b-4616-a80c-29a7e8e750e2","Type":"ContainerStarted","Data":"0f69434a9db9bceb12c66989fba99a7fce7624aaa43608837cd7f1d51a750edb"} Nov 28 15:46:13 crc kubenswrapper[4805]: I1128 15:46:13.358231 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" event={"ID":"a363e2d1-0a91-4578-b9e4-b736b7931b03","Type":"ContainerStarted","Data":"c1900187afec2dc4e77cef6a31189c5d8ad610ee80412555fe4dfc6358cea286"} Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.232441 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.232801 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.373411 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d99f7798d-mk9wl" event={"ID":"51176eca-5fa0-411d-976f-4095550240cd","Type":"ContainerStarted","Data":"9c63eb702150b3393258a9c8302ff4ec3b97d4bbbbf27d0d1583e32334575127"} Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.374048 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.374069 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.393383 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" event={"ID":"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950","Type":"ContainerStarted","Data":"f1236c3a6f401b280bc06c233c0efc5e15b6539effc24bdee4cab358f57dbb1c"} Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.394446 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.416011 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6d99f7798d-mk9wl" podStartSLOduration=3.415990579 podStartE2EDuration="3.415990579s" podCreationTimestamp="2025-11-28 15:46:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:14.405719569 +0000 UTC m=+1201.455510900" watchObservedRunningTime="2025-11-28 15:46:14.415990579 +0000 UTC m=+1201.465781900" Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.521477 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 15:46:14 crc kubenswrapper[4805]: I1128 15:46:14.551908 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" podStartSLOduration=3.551891256 podStartE2EDuration="3.551891256s" podCreationTimestamp="2025-11-28 15:46:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:14.429795363 +0000 UTC m=+1201.479586674" watchObservedRunningTime="2025-11-28 15:46:14.551891256 +0000 UTC m=+1201.601682567" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.105617 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5fc4b9b6bd-9nsgk"] Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.107388 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.109468 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.109500 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.127332 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fc4b9b6bd-9nsgk"] Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.209195 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.209525 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.209543 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-logs\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.209569 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-public-tls-certs\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.209725 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksfft\" (UniqueName: \"kubernetes.io/projected/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-kube-api-access-ksfft\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.209767 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-internal-tls-certs\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.209899 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-combined-ca-bundle\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.311062 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-internal-tls-certs\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.311114 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-combined-ca-bundle\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.311140 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.311163 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.311180 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-logs\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.311211 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-public-tls-certs\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.311310 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksfft\" (UniqueName: \"kubernetes.io/projected/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-kube-api-access-ksfft\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.313726 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-logs\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.319263 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-public-tls-certs\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.322818 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-internal-tls-certs\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.323757 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.323806 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-combined-ca-bundle\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.323926 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.334928 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksfft\" (UniqueName: \"kubernetes.io/projected/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-kube-api-access-ksfft\") pod \"barbican-api-5fc4b9b6bd-9nsgk\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.403868 4805 generic.go:334] "Generic (PLEG): container finished" podID="e013d7bd-1c72-4b87-be4b-83f895797124" containerID="7f4ba1f787d29d980bc0c209ec06ceece827e2ac14219237aa842b0f992266a1" exitCode=0 Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.403939 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zr49k" event={"ID":"e013d7bd-1c72-4b87-be4b-83f895797124","Type":"ContainerDied","Data":"7f4ba1f787d29d980bc0c209ec06ceece827e2ac14219237aa842b0f992266a1"} Nov 28 15:46:15 crc kubenswrapper[4805]: I1128 15:46:15.422770 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:21 crc kubenswrapper[4805]: I1128 15:46:21.924558 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:21 crc kubenswrapper[4805]: I1128 15:46:21.979169 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-np5d7"] Nov 28 15:46:21 crc kubenswrapper[4805]: I1128 15:46:21.979551 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f645789c-np5d7" podUID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerName="dnsmasq-dns" containerID="cri-o://851438c9ffd66c2d9c341bb6091b3cacc94b01f46299b3677347a2c85751bc4b" gracePeriod=10 Nov 28 15:46:23 crc kubenswrapper[4805]: I1128 15:46:23.556698 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:23 crc kubenswrapper[4805]: I1128 15:46:23.561701 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:24 crc kubenswrapper[4805]: I1128 15:46:24.305723 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f645789c-np5d7" podUID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: connect: connection refused" Nov 28 15:46:29 crc kubenswrapper[4805]: I1128 15:46:29.306466 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f645789c-np5d7" podUID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: connect: connection refused" Nov 28 15:46:29 crc kubenswrapper[4805]: I1128 15:46:29.540712 4805 generic.go:334] "Generic (PLEG): container finished" podID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerID="851438c9ffd66c2d9c341bb6091b3cacc94b01f46299b3677347a2c85751bc4b" exitCode=0 Nov 28 15:46:29 crc kubenswrapper[4805]: I1128 15:46:29.540798 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-np5d7" event={"ID":"ea1fc101-a095-4a2f-a0c3-9e13105a6077","Type":"ContainerDied","Data":"851438c9ffd66c2d9c341bb6091b3cacc94b01f46299b3677347a2c85751bc4b"} Nov 28 15:46:30 crc kubenswrapper[4805]: I1128 15:46:30.793469 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zr49k" Nov 28 15:46:30 crc kubenswrapper[4805]: I1128 15:46:30.927886 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qczgq\" (UniqueName: \"kubernetes.io/projected/e013d7bd-1c72-4b87-be4b-83f895797124-kube-api-access-qczgq\") pod \"e013d7bd-1c72-4b87-be4b-83f895797124\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " Nov 28 15:46:30 crc kubenswrapper[4805]: I1128 15:46:30.927970 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-config\") pod \"e013d7bd-1c72-4b87-be4b-83f895797124\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " Nov 28 15:46:30 crc kubenswrapper[4805]: I1128 15:46:30.928039 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-combined-ca-bundle\") pod \"e013d7bd-1c72-4b87-be4b-83f895797124\" (UID: \"e013d7bd-1c72-4b87-be4b-83f895797124\") " Nov 28 15:46:30 crc kubenswrapper[4805]: I1128 15:46:30.936589 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e013d7bd-1c72-4b87-be4b-83f895797124-kube-api-access-qczgq" (OuterVolumeSpecName: "kube-api-access-qczgq") pod "e013d7bd-1c72-4b87-be4b-83f895797124" (UID: "e013d7bd-1c72-4b87-be4b-83f895797124"). InnerVolumeSpecName "kube-api-access-qczgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:30 crc kubenswrapper[4805]: I1128 15:46:30.974066 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e013d7bd-1c72-4b87-be4b-83f895797124" (UID: "e013d7bd-1c72-4b87-be4b-83f895797124"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:30 crc kubenswrapper[4805]: I1128 15:46:30.992904 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-config" (OuterVolumeSpecName: "config") pod "e013d7bd-1c72-4b87-be4b-83f895797124" (UID: "e013d7bd-1c72-4b87-be4b-83f895797124"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.030286 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qczgq\" (UniqueName: \"kubernetes.io/projected/e013d7bd-1c72-4b87-be4b-83f895797124-kube-api-access-qczgq\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.030325 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.030339 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e013d7bd-1c72-4b87-be4b-83f895797124-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.560559 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zr49k" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.560487 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zr49k" event={"ID":"e013d7bd-1c72-4b87-be4b-83f895797124","Type":"ContainerDied","Data":"cd7ca3cb627038937bf5091115d44b0df30fd12144e13d061d3e5c763fd3a333"} Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.560711 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd7ca3cb627038937bf5091115d44b0df30fd12144e13d061d3e5c763fd3a333" Nov 28 15:46:31 crc kubenswrapper[4805]: E1128 15:46:31.589452 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:5849180ce2c77f49e1eecc517a821ce95b145e9702b347156f3b04da48eebe44" Nov 28 15:46:31 crc kubenswrapper[4805]: E1128 15:46:31.589602 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:5849180ce2c77f49e1eecc517a821ce95b145e9702b347156f3b04da48eebe44,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tvv7w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(122b157a-374d-437e-97bb-66fa032893ff): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:46:31 crc kubenswrapper[4805]: E1128 15:46:31.590943 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="122b157a-374d-437e-97bb-66fa032893ff" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.609554 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.747231 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-svc\") pod \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.747571 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-swift-storage-0\") pod \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.747598 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-sb\") pod \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.747660 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-nb\") pod \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.747754 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhvtn\" (UniqueName: \"kubernetes.io/projected/ea1fc101-a095-4a2f-a0c3-9e13105a6077-kube-api-access-jhvtn\") pod \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.747898 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-config\") pod \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.763755 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea1fc101-a095-4a2f-a0c3-9e13105a6077-kube-api-access-jhvtn" (OuterVolumeSpecName: "kube-api-access-jhvtn") pod "ea1fc101-a095-4a2f-a0c3-9e13105a6077" (UID: "ea1fc101-a095-4a2f-a0c3-9e13105a6077"). InnerVolumeSpecName "kube-api-access-jhvtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.796850 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea1fc101-a095-4a2f-a0c3-9e13105a6077" (UID: "ea1fc101-a095-4a2f-a0c3-9e13105a6077"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.850256 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-config" (OuterVolumeSpecName: "config") pod "ea1fc101-a095-4a2f-a0c3-9e13105a6077" (UID: "ea1fc101-a095-4a2f-a0c3-9e13105a6077"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.851073 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea1fc101-a095-4a2f-a0c3-9e13105a6077" (UID: "ea1fc101-a095-4a2f-a0c3-9e13105a6077"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.852640 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-config\") pod \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\" (UID: \"ea1fc101-a095-4a2f-a0c3-9e13105a6077\") " Nov 28 15:46:31 crc kubenswrapper[4805]: W1128 15:46:31.853281 4805 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/ea1fc101-a095-4a2f-a0c3-9e13105a6077/volumes/kubernetes.io~configmap/config Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.853346 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-config" (OuterVolumeSpecName: "config") pod "ea1fc101-a095-4a2f-a0c3-9e13105a6077" (UID: "ea1fc101-a095-4a2f-a0c3-9e13105a6077"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.858096 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea1fc101-a095-4a2f-a0c3-9e13105a6077" (UID: "ea1fc101-a095-4a2f-a0c3-9e13105a6077"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.863956 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.863988 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.863997 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.864006 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.864017 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhvtn\" (UniqueName: \"kubernetes.io/projected/ea1fc101-a095-4a2f-a0c3-9e13105a6077-kube-api-access-jhvtn\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.895970 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ea1fc101-a095-4a2f-a0c3-9e13105a6077" (UID: "ea1fc101-a095-4a2f-a0c3-9e13105a6077"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.967587 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea1fc101-a095-4a2f-a0c3-9e13105a6077-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.977889 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-rdcnt"] Nov 28 15:46:31 crc kubenswrapper[4805]: E1128 15:46:31.978345 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerName="dnsmasq-dns" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.978386 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerName="dnsmasq-dns" Nov 28 15:46:31 crc kubenswrapper[4805]: E1128 15:46:31.978405 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerName="init" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.978413 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerName="init" Nov 28 15:46:31 crc kubenswrapper[4805]: E1128 15:46:31.978449 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e013d7bd-1c72-4b87-be4b-83f895797124" containerName="neutron-db-sync" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.978466 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e013d7bd-1c72-4b87-be4b-83f895797124" containerName="neutron-db-sync" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.978658 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e013d7bd-1c72-4b87-be4b-83f895797124" containerName="neutron-db-sync" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.978708 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" containerName="dnsmasq-dns" Nov 28 15:46:31 crc kubenswrapper[4805]: I1128 15:46:31.979806 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.006991 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-rdcnt"] Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.041043 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fc4b9b6bd-9nsgk"] Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.069388 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxv9t\" (UniqueName: \"kubernetes.io/projected/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-kube-api-access-zxv9t\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.069465 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-nb\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.069522 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-sb\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.069543 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-swift-storage-0\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.069604 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-svc\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.069637 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-config\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.171262 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-config\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.171307 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxv9t\" (UniqueName: \"kubernetes.io/projected/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-kube-api-access-zxv9t\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.171369 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-nb\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.171419 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-sb\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.171438 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-swift-storage-0\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.171496 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-svc\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.172326 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-svc\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.172833 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-config\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.173552 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-nb\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.174029 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-sb\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.174794 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-swift-storage-0\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.184349 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-79b56c885b-5nx79"] Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.186125 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.200789 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dwvnp" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.203037 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.203282 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.209447 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.211311 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79b56c885b-5nx79"] Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.232849 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxv9t\" (UniqueName: \"kubernetes.io/projected/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-kube-api-access-zxv9t\") pod \"dnsmasq-dns-64dfd64c45-rdcnt\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.292270 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-ovndb-tls-certs\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.292326 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwm28\" (UniqueName: \"kubernetes.io/projected/764e5362-53f7-46e3-a31b-5a0c679f3484-kube-api-access-nwm28\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.292387 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-combined-ca-bundle\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.292459 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-httpd-config\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.292579 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-config\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.371119 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.394386 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwm28\" (UniqueName: \"kubernetes.io/projected/764e5362-53f7-46e3-a31b-5a0c679f3484-kube-api-access-nwm28\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.394437 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-ovndb-tls-certs\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.394467 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-combined-ca-bundle\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.394519 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-httpd-config\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.394602 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-config\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.402229 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-httpd-config\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.405163 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-combined-ca-bundle\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.406321 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-config\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.407038 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-ovndb-tls-certs\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.422098 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwm28\" (UniqueName: \"kubernetes.io/projected/764e5362-53f7-46e3-a31b-5a0c679f3484-kube-api-access-nwm28\") pod \"neutron-79b56c885b-5nx79\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.459918 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.596895 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" event={"ID":"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6","Type":"ContainerStarted","Data":"e4e089980980ac67dc67f214af92d43970eb148b6a828b8e14ba26fab9d1df86"} Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.597259 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" event={"ID":"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6","Type":"ContainerStarted","Data":"bde01a0dd31be7ceea17360d55a6773739af8157b4ce25586760b7f62af80a26"} Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.604791 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" event={"ID":"a363e2d1-0a91-4578-b9e4-b736b7931b03","Type":"ContainerStarted","Data":"66fc00b5a0c7d5310a8c3fbb5a766ab2e30e4c1049270d558c221c8c94d6292e"} Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.619785 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-np5d7" event={"ID":"ea1fc101-a095-4a2f-a0c3-9e13105a6077","Type":"ContainerDied","Data":"0df66dd49ed31efd381ea52d0e6e91bd79b8cf082aba66f5095a3a8dcfccb00c"} Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.619837 4805 scope.go:117] "RemoveContainer" containerID="851438c9ffd66c2d9c341bb6091b3cacc94b01f46299b3677347a2c85751bc4b" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.619994 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f645789c-np5d7" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.669139 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-97ccb4d45-ddn2r" event={"ID":"abfd499c-e83b-4616-a80c-29a7e8e750e2","Type":"ContainerStarted","Data":"c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5"} Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.669236 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="ceilometer-central-agent" containerID="cri-o://212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66" gracePeriod=30 Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.669348 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="sg-core" containerID="cri-o://a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43" gracePeriod=30 Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.669418 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="ceilometer-notification-agent" containerID="cri-o://4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd" gracePeriod=30 Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.724557 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-np5d7"] Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.739752 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-np5d7"] Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.773766 4805 scope.go:117] "RemoveContainer" containerID="138024a909782c1f5bbc8a05dcb191e646e71777eca84bbe32c86ae23a29abab" Nov 28 15:46:32 crc kubenswrapper[4805]: I1128 15:46:32.944352 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-rdcnt"] Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.221321 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea1fc101-a095-4a2f-a0c3-9e13105a6077" path="/var/lib/kubelet/pods/ea1fc101-a095-4a2f-a0c3-9e13105a6077/volumes" Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.299199 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79b56c885b-5nx79"] Nov 28 15:46:33 crc kubenswrapper[4805]: W1128 15:46:33.330608 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod764e5362_53f7_46e3_a31b_5a0c679f3484.slice/crio-7e998d046840d741e1be77452ff2eba6aeae6eb242cecd212ab0c168441e4b42 WatchSource:0}: Error finding container 7e998d046840d741e1be77452ff2eba6aeae6eb242cecd212ab0c168441e4b42: Status 404 returned error can't find the container with id 7e998d046840d741e1be77452ff2eba6aeae6eb242cecd212ab0c168441e4b42 Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.682384 4805 generic.go:334] "Generic (PLEG): container finished" podID="122b157a-374d-437e-97bb-66fa032893ff" containerID="a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43" exitCode=2 Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.682424 4805 generic.go:334] "Generic (PLEG): container finished" podID="122b157a-374d-437e-97bb-66fa032893ff" containerID="212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66" exitCode=0 Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.682476 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"122b157a-374d-437e-97bb-66fa032893ff","Type":"ContainerDied","Data":"a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.682510 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"122b157a-374d-437e-97bb-66fa032893ff","Type":"ContainerDied","Data":"212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.685618 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-97ccb4d45-ddn2r" event={"ID":"abfd499c-e83b-4616-a80c-29a7e8e750e2","Type":"ContainerStarted","Data":"1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.688641 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b56c885b-5nx79" event={"ID":"764e5362-53f7-46e3-a31b-5a0c679f3484","Type":"ContainerStarted","Data":"904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.688681 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b56c885b-5nx79" event={"ID":"764e5362-53f7-46e3-a31b-5a0c679f3484","Type":"ContainerStarted","Data":"7e998d046840d741e1be77452ff2eba6aeae6eb242cecd212ab0c168441e4b42"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.690528 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" event={"ID":"a363e2d1-0a91-4578-b9e4-b736b7931b03","Type":"ContainerStarted","Data":"02772248390f9aa46863e063f9eac9961344ec1704c67c8cfde40db99e473634"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.693303 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" event={"ID":"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6","Type":"ContainerStarted","Data":"42eba8d003a7a3309a588e3a7f4e6e3ccf9656400a9603ef4fb2b81bd3c138d4"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.693392 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.695070 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zwp6n" event={"ID":"b2f218fa-9d63-46f4-ba0d-3b9166d23672","Type":"ContainerStarted","Data":"65df2285b683dab8a99695a41df58b3ed9b436df295bd383c9af5720935f1dbb"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.697185 4805 generic.go:334] "Generic (PLEG): container finished" podID="9e3a63a9-3d98-4b2a-addb-ed8db9817da8" containerID="d61288754ef3eb1a5f37e9e9dacb08c17105086bfb4532fc0d9c7d82e4c8e60d" exitCode=0 Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.697219 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" event={"ID":"9e3a63a9-3d98-4b2a-addb-ed8db9817da8","Type":"ContainerDied","Data":"d61288754ef3eb1a5f37e9e9dacb08c17105086bfb4532fc0d9c7d82e4c8e60d"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.697235 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" event={"ID":"9e3a63a9-3d98-4b2a-addb-ed8db9817da8","Type":"ContainerStarted","Data":"279c32d43ba009f747c3cd132f3462abed9cd5857b2dc11a108d69ad06964b3a"} Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.707538 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-97ccb4d45-ddn2r" podStartSLOduration=3.456879521 podStartE2EDuration="22.70750627s" podCreationTimestamp="2025-11-28 15:46:11 +0000 UTC" firstStartedPulling="2025-11-28 15:46:12.36884798 +0000 UTC m=+1199.418639291" lastFinishedPulling="2025-11-28 15:46:31.619474719 +0000 UTC m=+1218.669266040" observedRunningTime="2025-11-28 15:46:33.705611228 +0000 UTC m=+1220.755402549" watchObservedRunningTime="2025-11-28 15:46:33.70750627 +0000 UTC m=+1220.757297591" Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.735519 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" podStartSLOduration=3.644227159 podStartE2EDuration="22.735503362s" podCreationTimestamp="2025-11-28 15:46:11 +0000 UTC" firstStartedPulling="2025-11-28 15:46:12.523808777 +0000 UTC m=+1199.573600088" lastFinishedPulling="2025-11-28 15:46:31.61508498 +0000 UTC m=+1218.664876291" observedRunningTime="2025-11-28 15:46:33.724980315 +0000 UTC m=+1220.774771626" watchObservedRunningTime="2025-11-28 15:46:33.735503362 +0000 UTC m=+1220.785294673" Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.753782 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-zwp6n" podStartSLOduration=5.373983639 podStartE2EDuration="1m5.753763178s" podCreationTimestamp="2025-11-28 15:45:28 +0000 UTC" firstStartedPulling="2025-11-28 15:45:30.328697049 +0000 UTC m=+1157.378488360" lastFinishedPulling="2025-11-28 15:46:30.708476598 +0000 UTC m=+1217.758267899" observedRunningTime="2025-11-28 15:46:33.744439525 +0000 UTC m=+1220.794230836" watchObservedRunningTime="2025-11-28 15:46:33.753763178 +0000 UTC m=+1220.803554479" Nov 28 15:46:33 crc kubenswrapper[4805]: I1128 15:46:33.776164 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" podStartSLOduration=18.776148347 podStartE2EDuration="18.776148347s" podCreationTimestamp="2025-11-28 15:46:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:33.767544223 +0000 UTC m=+1220.817335544" watchObservedRunningTime="2025-11-28 15:46:33.776148347 +0000 UTC m=+1220.825939658" Nov 28 15:46:34 crc kubenswrapper[4805]: I1128 15:46:34.738612 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b56c885b-5nx79" event={"ID":"764e5362-53f7-46e3-a31b-5a0c679f3484","Type":"ContainerStarted","Data":"2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee"} Nov 28 15:46:34 crc kubenswrapper[4805]: I1128 15:46:34.740081 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:46:34 crc kubenswrapper[4805]: I1128 15:46:34.758786 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" event={"ID":"9e3a63a9-3d98-4b2a-addb-ed8db9817da8","Type":"ContainerStarted","Data":"7b57209f68f5d6cc8684f9998bd64f6a82b76326c7b7ad217f18f3807664a4c7"} Nov 28 15:46:34 crc kubenswrapper[4805]: I1128 15:46:34.758826 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:34 crc kubenswrapper[4805]: I1128 15:46:34.759722 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:34 crc kubenswrapper[4805]: I1128 15:46:34.785649 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-79b56c885b-5nx79" podStartSLOduration=2.785631769 podStartE2EDuration="2.785631769s" podCreationTimestamp="2025-11-28 15:46:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:34.781728302 +0000 UTC m=+1221.831519613" watchObservedRunningTime="2025-11-28 15:46:34.785631769 +0000 UTC m=+1221.835423070" Nov 28 15:46:34 crc kubenswrapper[4805]: I1128 15:46:34.834842 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" podStartSLOduration=3.834824577 podStartE2EDuration="3.834824577s" podCreationTimestamp="2025-11-28 15:46:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:34.824963799 +0000 UTC m=+1221.874755110" watchObservedRunningTime="2025-11-28 15:46:34.834824577 +0000 UTC m=+1221.884615888" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.235627 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6bdfcdc49-v4hgd"] Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.237376 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.238943 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.240506 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.250858 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bdfcdc49-v4hgd"] Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.361160 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cngjl\" (UniqueName: \"kubernetes.io/projected/8724692d-29b7-46c8-b0dd-802e570050a7-kube-api-access-cngjl\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.361249 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-ovndb-tls-certs\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.361292 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-public-tls-certs\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.361330 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-httpd-config\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.361408 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-internal-tls-certs\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.361431 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-config\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.361453 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-combined-ca-bundle\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.462803 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-ovndb-tls-certs\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.462891 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-public-tls-certs\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.462947 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-httpd-config\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.462999 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-internal-tls-certs\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.463029 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-config\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.463055 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-combined-ca-bundle\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.463149 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cngjl\" (UniqueName: \"kubernetes.io/projected/8724692d-29b7-46c8-b0dd-802e570050a7-kube-api-access-cngjl\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.470347 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-public-tls-certs\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.471345 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-internal-tls-certs\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.473062 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-httpd-config\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.473592 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-ovndb-tls-certs\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.475804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-config\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.478248 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-combined-ca-bundle\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.481759 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cngjl\" (UniqueName: \"kubernetes.io/projected/8724692d-29b7-46c8-b0dd-802e570050a7-kube-api-access-cngjl\") pod \"neutron-6bdfcdc49-v4hgd\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:35 crc kubenswrapper[4805]: I1128 15:46:35.559324 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.206778 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bdfcdc49-v4hgd"] Nov 28 15:46:36 crc kubenswrapper[4805]: W1128 15:46:36.208158 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8724692d_29b7_46c8_b0dd_802e570050a7.slice/crio-6031a39a57464feb543c3973c7de58f3b2aac821b73b878c3d11850ed778203c WatchSource:0}: Error finding container 6031a39a57464feb543c3973c7de58f3b2aac821b73b878c3d11850ed778203c: Status 404 returned error can't find the container with id 6031a39a57464feb543c3973c7de58f3b2aac821b73b878c3d11850ed778203c Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.251971 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.380268 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-combined-ca-bundle\") pod \"122b157a-374d-437e-97bb-66fa032893ff\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.380648 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-run-httpd\") pod \"122b157a-374d-437e-97bb-66fa032893ff\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.380697 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-config-data\") pod \"122b157a-374d-437e-97bb-66fa032893ff\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.380773 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-scripts\") pod \"122b157a-374d-437e-97bb-66fa032893ff\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.380816 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvv7w\" (UniqueName: \"kubernetes.io/projected/122b157a-374d-437e-97bb-66fa032893ff-kube-api-access-tvv7w\") pod \"122b157a-374d-437e-97bb-66fa032893ff\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.380837 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-sg-core-conf-yaml\") pod \"122b157a-374d-437e-97bb-66fa032893ff\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.380875 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-log-httpd\") pod \"122b157a-374d-437e-97bb-66fa032893ff\" (UID: \"122b157a-374d-437e-97bb-66fa032893ff\") " Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.381526 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "122b157a-374d-437e-97bb-66fa032893ff" (UID: "122b157a-374d-437e-97bb-66fa032893ff"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.381663 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "122b157a-374d-437e-97bb-66fa032893ff" (UID: "122b157a-374d-437e-97bb-66fa032893ff"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.390294 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-scripts" (OuterVolumeSpecName: "scripts") pod "122b157a-374d-437e-97bb-66fa032893ff" (UID: "122b157a-374d-437e-97bb-66fa032893ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.390379 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/122b157a-374d-437e-97bb-66fa032893ff-kube-api-access-tvv7w" (OuterVolumeSpecName: "kube-api-access-tvv7w") pod "122b157a-374d-437e-97bb-66fa032893ff" (UID: "122b157a-374d-437e-97bb-66fa032893ff"). InnerVolumeSpecName "kube-api-access-tvv7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.427213 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "122b157a-374d-437e-97bb-66fa032893ff" (UID: "122b157a-374d-437e-97bb-66fa032893ff"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.436428 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-config-data" (OuterVolumeSpecName: "config-data") pod "122b157a-374d-437e-97bb-66fa032893ff" (UID: "122b157a-374d-437e-97bb-66fa032893ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.447024 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "122b157a-374d-437e-97bb-66fa032893ff" (UID: "122b157a-374d-437e-97bb-66fa032893ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.482628 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.482670 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvv7w\" (UniqueName: \"kubernetes.io/projected/122b157a-374d-437e-97bb-66fa032893ff-kube-api-access-tvv7w\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.482684 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.482696 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.482737 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/122b157a-374d-437e-97bb-66fa032893ff-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.482751 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.482761 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122b157a-374d-437e-97bb-66fa032893ff-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.778876 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdfcdc49-v4hgd" event={"ID":"8724692d-29b7-46c8-b0dd-802e570050a7","Type":"ContainerStarted","Data":"c253b5bf7e841bb27ed1346fc3c401ce938ec4f3dce125f741a8913badce2e99"} Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.778923 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdfcdc49-v4hgd" event={"ID":"8724692d-29b7-46c8-b0dd-802e570050a7","Type":"ContainerStarted","Data":"7569e16b1b6d818541ba1419b8d0036700d58b74a89a37e7bc67e8d06bfaeb15"} Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.778938 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdfcdc49-v4hgd" event={"ID":"8724692d-29b7-46c8-b0dd-802e570050a7","Type":"ContainerStarted","Data":"6031a39a57464feb543c3973c7de58f3b2aac821b73b878c3d11850ed778203c"} Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.780154 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.782236 4805 generic.go:334] "Generic (PLEG): container finished" podID="122b157a-374d-437e-97bb-66fa032893ff" containerID="4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd" exitCode=0 Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.782978 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.787848 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"122b157a-374d-437e-97bb-66fa032893ff","Type":"ContainerDied","Data":"4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd"} Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.787901 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"122b157a-374d-437e-97bb-66fa032893ff","Type":"ContainerDied","Data":"bb341155e49ee01dfea40123ea6e5bad22b83e1cc82b5f1492d7f6cfb18dfd41"} Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.787925 4805 scope.go:117] "RemoveContainer" containerID="a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.844192 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6bdfcdc49-v4hgd" podStartSLOduration=1.8441686069999998 podStartE2EDuration="1.844168607s" podCreationTimestamp="2025-11-28 15:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:36.805238827 +0000 UTC m=+1223.855030138" watchObservedRunningTime="2025-11-28 15:46:36.844168607 +0000 UTC m=+1223.893959928" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.923525 4805 scope.go:117] "RemoveContainer" containerID="4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.936644 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.952821 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.967607 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:36 crc kubenswrapper[4805]: E1128 15:46:36.968573 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="ceilometer-central-agent" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.968595 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="ceilometer-central-agent" Nov 28 15:46:36 crc kubenswrapper[4805]: E1128 15:46:36.968639 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="ceilometer-notification-agent" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.968648 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="ceilometer-notification-agent" Nov 28 15:46:36 crc kubenswrapper[4805]: E1128 15:46:36.968657 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="sg-core" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.968663 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="sg-core" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.968971 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="ceilometer-notification-agent" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.969550 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="sg-core" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.969580 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="122b157a-374d-437e-97bb-66fa032893ff" containerName="ceilometer-central-agent" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.974095 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.974747 4805 scope.go:117] "RemoveContainer" containerID="212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.978702 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.979590 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 15:46:36 crc kubenswrapper[4805]: I1128 15:46:36.980657 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.011595 4805 scope.go:117] "RemoveContainer" containerID="a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43" Nov 28 15:46:37 crc kubenswrapper[4805]: E1128 15:46:37.013845 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43\": container with ID starting with a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43 not found: ID does not exist" containerID="a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.013895 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43"} err="failed to get container status \"a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43\": rpc error: code = NotFound desc = could not find container \"a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43\": container with ID starting with a7e844d58f5f2fee661956331ab39e88e4128fbeaed4a5ea6989021550ee9c43 not found: ID does not exist" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.013937 4805 scope.go:117] "RemoveContainer" containerID="4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd" Nov 28 15:46:37 crc kubenswrapper[4805]: E1128 15:46:37.024509 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd\": container with ID starting with 4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd not found: ID does not exist" containerID="4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.024557 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd"} err="failed to get container status \"4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd\": rpc error: code = NotFound desc = could not find container \"4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd\": container with ID starting with 4f46a8a2dc9d0f07b2a66bd8f155ca6c3eb3f515f79ee07e5ad4f8642deef1cd not found: ID does not exist" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.024587 4805 scope.go:117] "RemoveContainer" containerID="212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66" Nov 28 15:46:37 crc kubenswrapper[4805]: E1128 15:46:37.025111 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66\": container with ID starting with 212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66 not found: ID does not exist" containerID="212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.025234 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66"} err="failed to get container status \"212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66\": rpc error: code = NotFound desc = could not find container \"212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66\": container with ID starting with 212726f3c7ff7be55104ed086d4c231c1c9684db153fe562be2222c10b6e0b66 not found: ID does not exist" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.135588 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-run-httpd\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.136797 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.137041 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-log-httpd\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.137283 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-scripts\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.137544 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-config-data\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.137708 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg5px\" (UniqueName: \"kubernetes.io/projected/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-kube-api-access-qg5px\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.137889 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.219534 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="122b157a-374d-437e-97bb-66fa032893ff" path="/var/lib/kubelet/pods/122b157a-374d-437e-97bb-66fa032893ff/volumes" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.240286 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-scripts\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.240400 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-config-data\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.240436 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg5px\" (UniqueName: \"kubernetes.io/projected/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-kube-api-access-qg5px\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.240510 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.240627 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-run-httpd\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.240659 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.240725 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-log-httpd\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.241455 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-log-httpd\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.242848 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-run-httpd\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.249840 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-scripts\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.251350 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.253786 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.254894 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-config-data\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.264824 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg5px\" (UniqueName: \"kubernetes.io/projected/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-kube-api-access-qg5px\") pod \"ceilometer-0\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.300014 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:46:37 crc kubenswrapper[4805]: I1128 15:46:37.890447 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:38 crc kubenswrapper[4805]: I1128 15:46:38.810819 4805 generic.go:334] "Generic (PLEG): container finished" podID="b2f218fa-9d63-46f4-ba0d-3b9166d23672" containerID="65df2285b683dab8a99695a41df58b3ed9b436df295bd383c9af5720935f1dbb" exitCode=0 Nov 28 15:46:38 crc kubenswrapper[4805]: I1128 15:46:38.810906 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zwp6n" event={"ID":"b2f218fa-9d63-46f4-ba0d-3b9166d23672","Type":"ContainerDied","Data":"65df2285b683dab8a99695a41df58b3ed9b436df295bd383c9af5720935f1dbb"} Nov 28 15:46:38 crc kubenswrapper[4805]: I1128 15:46:38.812688 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerStarted","Data":"272aa407188898c5f5222aa4d11cbe180785d3c06917d009a59e004294569021"} Nov 28 15:46:38 crc kubenswrapper[4805]: I1128 15:46:38.812732 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerStarted","Data":"5ab4c0d7a7df62739992e348d107282bddc8eb0fe733dd64e5d70ed878ff4494"} Nov 28 15:46:39 crc kubenswrapper[4805]: I1128 15:46:39.807415 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:39 crc kubenswrapper[4805]: I1128 15:46:39.812328 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:46:39 crc kubenswrapper[4805]: I1128 15:46:39.840086 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerStarted","Data":"cbd1edd2530ac67b1f84c5e72a6cca109f3a42bf326efdf16622f58cb57233e3"} Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.451861 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.620178 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bvd5\" (UniqueName: \"kubernetes.io/projected/b2f218fa-9d63-46f4-ba0d-3b9166d23672-kube-api-access-9bvd5\") pod \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.620255 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-config-data\") pod \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.620321 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-scripts\") pod \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.620431 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b2f218fa-9d63-46f4-ba0d-3b9166d23672-etc-machine-id\") pod \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.620478 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-db-sync-config-data\") pod \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.620525 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-combined-ca-bundle\") pod \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\" (UID: \"b2f218fa-9d63-46f4-ba0d-3b9166d23672\") " Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.625210 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2f218fa-9d63-46f4-ba0d-3b9166d23672-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b2f218fa-9d63-46f4-ba0d-3b9166d23672" (UID: "b2f218fa-9d63-46f4-ba0d-3b9166d23672"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.626643 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-scripts" (OuterVolumeSpecName: "scripts") pod "b2f218fa-9d63-46f4-ba0d-3b9166d23672" (UID: "b2f218fa-9d63-46f4-ba0d-3b9166d23672"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.629651 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2f218fa-9d63-46f4-ba0d-3b9166d23672-kube-api-access-9bvd5" (OuterVolumeSpecName: "kube-api-access-9bvd5") pod "b2f218fa-9d63-46f4-ba0d-3b9166d23672" (UID: "b2f218fa-9d63-46f4-ba0d-3b9166d23672"). InnerVolumeSpecName "kube-api-access-9bvd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.636546 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b2f218fa-9d63-46f4-ba0d-3b9166d23672" (UID: "b2f218fa-9d63-46f4-ba0d-3b9166d23672"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.655326 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.665487 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2f218fa-9d63-46f4-ba0d-3b9166d23672" (UID: "b2f218fa-9d63-46f4-ba0d-3b9166d23672"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.715286 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-config-data" (OuterVolumeSpecName: "config-data") pod "b2f218fa-9d63-46f4-ba0d-3b9166d23672" (UID: "b2f218fa-9d63-46f4-ba0d-3b9166d23672"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.723007 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bvd5\" (UniqueName: \"kubernetes.io/projected/b2f218fa-9d63-46f4-ba0d-3b9166d23672-kube-api-access-9bvd5\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.723054 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.723069 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.723084 4805 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b2f218fa-9d63-46f4-ba0d-3b9166d23672-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.723096 4805 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.723107 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f218fa-9d63-46f4-ba0d-3b9166d23672-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.894090 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zwp6n" event={"ID":"b2f218fa-9d63-46f4-ba0d-3b9166d23672","Type":"ContainerDied","Data":"f2a51e83a816cbb3a6cc732bf2fc8ded675b42b1c3b3cb3176308ccc3ab2c1f4"} Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.894138 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2a51e83a816cbb3a6cc732bf2fc8ded675b42b1c3b3cb3176308ccc3ab2c1f4" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.894226 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zwp6n" Nov 28 15:46:40 crc kubenswrapper[4805]: I1128 15:46:40.926006 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerStarted","Data":"62be3d9ee2c288accb4e7e1c932b1555386148b92540d70e2a4b9211551ee385"} Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.101541 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:46:41 crc kubenswrapper[4805]: E1128 15:46:41.102038 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2f218fa-9d63-46f4-ba0d-3b9166d23672" containerName="cinder-db-sync" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.102065 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2f218fa-9d63-46f4-ba0d-3b9166d23672" containerName="cinder-db-sync" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.102291 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2f218fa-9d63-46f4-ba0d-3b9166d23672" containerName="cinder-db-sync" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.103529 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.109970 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.110207 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gddf4" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.110902 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.111064 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.177506 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.242414 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.242488 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.242578 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-scripts\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.242622 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.242645 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43372220-e57d-45be-a9ec-ff7cdc0f5e02-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.242672 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8zs5\" (UniqueName: \"kubernetes.io/projected/43372220-e57d-45be-a9ec-ff7cdc0f5e02-kube-api-access-d8zs5\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.295480 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-rdcnt"] Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.298416 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" podUID="9e3a63a9-3d98-4b2a-addb-ed8db9817da8" containerName="dnsmasq-dns" containerID="cri-o://7b57209f68f5d6cc8684f9998bd64f6a82b76326c7b7ad217f18f3807664a4c7" gracePeriod=10 Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.305061 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-mdsn8"] Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.306600 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.306899 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.324460 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-mdsn8"] Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.348281 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-scripts\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.348658 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.348686 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43372220-e57d-45be-a9ec-ff7cdc0f5e02-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.348725 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8zs5\" (UniqueName: \"kubernetes.io/projected/43372220-e57d-45be-a9ec-ff7cdc0f5e02-kube-api-access-d8zs5\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.348762 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.348803 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.349609 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43372220-e57d-45be-a9ec-ff7cdc0f5e02-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.373058 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.400625 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.412223 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.417336 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.420881 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-scripts\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.430191 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.434912 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.446862 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8zs5\" (UniqueName: \"kubernetes.io/projected/43372220-e57d-45be-a9ec-ff7cdc0f5e02-kube-api-access-d8zs5\") pod \"cinder-scheduler-0\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.454267 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-config\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.454309 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-sb\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.468080 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.479606 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-svc\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.479776 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-swift-storage-0\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.479830 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-nb\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.479876 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvjhx\" (UniqueName: \"kubernetes.io/projected/b77f479b-8072-4e56-8517-cce5ec74774c-kube-api-access-wvjhx\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.483917 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582528 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-swift-storage-0\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582598 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-nb\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582628 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-logs\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582666 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvjhx\" (UniqueName: \"kubernetes.io/projected/b77f479b-8072-4e56-8517-cce5ec74774c-kube-api-access-wvjhx\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582735 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data-custom\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582779 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-config\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582803 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-sb\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582837 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582859 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvxbd\" (UniqueName: \"kubernetes.io/projected/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-kube-api-access-lvxbd\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582895 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-scripts\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582933 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582960 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-svc\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.582983 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.583963 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-nb\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.584196 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-swift-storage-0\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.584775 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-config\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.584994 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-svc\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.588934 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-sb\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.620087 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvjhx\" (UniqueName: \"kubernetes.io/projected/b77f479b-8072-4e56-8517-cce5ec74774c-kube-api-access-wvjhx\") pod \"dnsmasq-dns-8ccb5c7cf-mdsn8\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.652866 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.686866 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-scripts\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.686909 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.686933 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.686984 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-logs\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.687035 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data-custom\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.687070 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.687085 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvxbd\" (UniqueName: \"kubernetes.io/projected/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-kube-api-access-lvxbd\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.688671 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-logs\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.688773 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.696495 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.701188 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data-custom\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.702582 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-scripts\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.710299 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvxbd\" (UniqueName: \"kubernetes.io/projected/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-kube-api-access-lvxbd\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.721972 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data\") pod \"cinder-api-0\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.920844 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.965050 4805 generic.go:334] "Generic (PLEG): container finished" podID="9e3a63a9-3d98-4b2a-addb-ed8db9817da8" containerID="7b57209f68f5d6cc8684f9998bd64f6a82b76326c7b7ad217f18f3807664a4c7" exitCode=0 Nov 28 15:46:41 crc kubenswrapper[4805]: I1128 15:46:41.965550 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" event={"ID":"9e3a63a9-3d98-4b2a-addb-ed8db9817da8","Type":"ContainerDied","Data":"7b57209f68f5d6cc8684f9998bd64f6a82b76326c7b7ad217f18f3807664a4c7"} Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.072014 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.098526 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-nb\") pod \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.098743 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxv9t\" (UniqueName: \"kubernetes.io/projected/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-kube-api-access-zxv9t\") pod \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.098886 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-swift-storage-0\") pod \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.099051 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-sb\") pod \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.099150 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-svc\") pod \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.099221 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-config\") pod \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\" (UID: \"9e3a63a9-3d98-4b2a-addb-ed8db9817da8\") " Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.132987 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-kube-api-access-zxv9t" (OuterVolumeSpecName: "kube-api-access-zxv9t") pod "9e3a63a9-3d98-4b2a-addb-ed8db9817da8" (UID: "9e3a63a9-3d98-4b2a-addb-ed8db9817da8"). InnerVolumeSpecName "kube-api-access-zxv9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.180574 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.203399 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxv9t\" (UniqueName: \"kubernetes.io/projected/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-kube-api-access-zxv9t\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.270271 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9e3a63a9-3d98-4b2a-addb-ed8db9817da8" (UID: "9e3a63a9-3d98-4b2a-addb-ed8db9817da8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.287659 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-config" (OuterVolumeSpecName: "config") pod "9e3a63a9-3d98-4b2a-addb-ed8db9817da8" (UID: "9e3a63a9-3d98-4b2a-addb-ed8db9817da8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.310777 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.310803 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.311840 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e3a63a9-3d98-4b2a-addb-ed8db9817da8" (UID: "9e3a63a9-3d98-4b2a-addb-ed8db9817da8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.321865 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e3a63a9-3d98-4b2a-addb-ed8db9817da8" (UID: "9e3a63a9-3d98-4b2a-addb-ed8db9817da8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.339505 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9e3a63a9-3d98-4b2a-addb-ed8db9817da8" (UID: "9e3a63a9-3d98-4b2a-addb-ed8db9817da8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.412093 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.412127 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.412136 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3a63a9-3d98-4b2a-addb-ed8db9817da8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.436842 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-mdsn8"] Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.708203 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.990419 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43372220-e57d-45be-a9ec-ff7cdc0f5e02","Type":"ContainerStarted","Data":"ed49a51280298e3afa79ad05b7d940e4933a67c330bb4f37172ee83f9747dc5a"} Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.994502 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" event={"ID":"9e3a63a9-3d98-4b2a-addb-ed8db9817da8","Type":"ContainerDied","Data":"279c32d43ba009f747c3cd132f3462abed9cd5857b2dc11a108d69ad06964b3a"} Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.994565 4805 scope.go:117] "RemoveContainer" containerID="7b57209f68f5d6cc8684f9998bd64f6a82b76326c7b7ad217f18f3807664a4c7" Nov 28 15:46:42 crc kubenswrapper[4805]: I1128 15:46:42.994787 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dfd64c45-rdcnt" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.002037 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e","Type":"ContainerStarted","Data":"2d30f3d6cf7ca0a747607cee472eb2d1cd713d3a8be1b4b1ae3a98d8921d9fe8"} Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.017378 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" event={"ID":"b77f479b-8072-4e56-8517-cce5ec74774c","Type":"ContainerStarted","Data":"8822aa83abea6917bb9adf5f7a52fe9ccb62d1e60afabde9ce668bb84df84ecd"} Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.043487 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-rdcnt"] Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.062289 4805 scope.go:117] "RemoveContainer" containerID="d61288754ef3eb1a5f37e9e9dacb08c17105086bfb4532fc0d9c7d82e4c8e60d" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.095436 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-rdcnt"] Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.111438 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 15:46:43 crc kubenswrapper[4805]: E1128 15:46:43.112025 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e3a63a9-3d98-4b2a-addb-ed8db9817da8" containerName="init" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.112104 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e3a63a9-3d98-4b2a-addb-ed8db9817da8" containerName="init" Nov 28 15:46:43 crc kubenswrapper[4805]: E1128 15:46:43.112191 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e3a63a9-3d98-4b2a-addb-ed8db9817da8" containerName="dnsmasq-dns" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.112275 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e3a63a9-3d98-4b2a-addb-ed8db9817da8" containerName="dnsmasq-dns" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.112544 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e3a63a9-3d98-4b2a-addb-ed8db9817da8" containerName="dnsmasq-dns" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.113215 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.116215 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.117759 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.117819 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-b4rj8" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.117971 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.236348 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.236579 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config-secret\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.236639 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnlnl\" (UniqueName: \"kubernetes.io/projected/f44dcea2-910c-462b-b3a1-785f93650002-kube-api-access-nnlnl\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.236658 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.261712 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e3a63a9-3d98-4b2a-addb-ed8db9817da8" path="/var/lib/kubelet/pods/9e3a63a9-3d98-4b2a-addb-ed8db9817da8/volumes" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.272375 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.338436 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config-secret\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.338515 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnlnl\" (UniqueName: \"kubernetes.io/projected/f44dcea2-910c-462b-b3a1-785f93650002-kube-api-access-nnlnl\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.338540 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.338629 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.340753 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.343853 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config-secret\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.352633 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.361856 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnlnl\" (UniqueName: \"kubernetes.io/projected/f44dcea2-910c-462b-b3a1-785f93650002-kube-api-access-nnlnl\") pod \"openstackclient\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.394350 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.395291 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.420060 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.443572 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.444803 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.455129 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.543797 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcggf\" (UniqueName: \"kubernetes.io/projected/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-kube-api-access-gcggf\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.543891 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.543927 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config-secret\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.543962 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: E1128 15:46:43.617750 4805 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 28 15:46:43 crc kubenswrapper[4805]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_f44dcea2-910c-462b-b3a1-785f93650002_0(224f1e33a747533d508cf4511c47994c60f36218483becd1354ac8605494e212): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"224f1e33a747533d508cf4511c47994c60f36218483becd1354ac8605494e212" Netns:"/var/run/netns/a56e37f7-b95b-4d50-bdf4-e987da9de932" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=224f1e33a747533d508cf4511c47994c60f36218483becd1354ac8605494e212;K8S_POD_UID=f44dcea2-910c-462b-b3a1-785f93650002" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/f44dcea2-910c-462b-b3a1-785f93650002]: expected pod UID "f44dcea2-910c-462b-b3a1-785f93650002" but got "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" from Kube API Nov 28 15:46:43 crc kubenswrapper[4805]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 28 15:46:43 crc kubenswrapper[4805]: > Nov 28 15:46:43 crc kubenswrapper[4805]: E1128 15:46:43.617842 4805 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 28 15:46:43 crc kubenswrapper[4805]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_f44dcea2-910c-462b-b3a1-785f93650002_0(224f1e33a747533d508cf4511c47994c60f36218483becd1354ac8605494e212): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"224f1e33a747533d508cf4511c47994c60f36218483becd1354ac8605494e212" Netns:"/var/run/netns/a56e37f7-b95b-4d50-bdf4-e987da9de932" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=224f1e33a747533d508cf4511c47994c60f36218483becd1354ac8605494e212;K8S_POD_UID=f44dcea2-910c-462b-b3a1-785f93650002" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/f44dcea2-910c-462b-b3a1-785f93650002]: expected pod UID "f44dcea2-910c-462b-b3a1-785f93650002" but got "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" from Kube API Nov 28 15:46:43 crc kubenswrapper[4805]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 28 15:46:43 crc kubenswrapper[4805]: > pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.647482 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcggf\" (UniqueName: \"kubernetes.io/projected/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-kube-api-access-gcggf\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.647906 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.647950 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config-secret\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.648033 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.649156 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.684942 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config-secret\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.685214 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.691008 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcggf\" (UniqueName: \"kubernetes.io/projected/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-kube-api-access-gcggf\") pod \"openstackclient\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " pod="openstack/openstackclient" Nov 28 15:46:43 crc kubenswrapper[4805]: I1128 15:46:43.786741 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.049698 4805 generic.go:334] "Generic (PLEG): container finished" podID="b77f479b-8072-4e56-8517-cce5ec74774c" containerID="bd5a19118f4db95a9c973e64adabb9e0dc85e4fbbc7a09353658f5767597a719" exitCode=0 Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.050019 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" event={"ID":"b77f479b-8072-4e56-8517-cce5ec74774c","Type":"ContainerDied","Data":"bd5a19118f4db95a9c973e64adabb9e0dc85e4fbbc7a09353658f5767597a719"} Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.081548 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerStarted","Data":"b842b88b2bff9eb8270ea244a59e110d63b3fff409cac01ed749dc5086362891"} Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.091615 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.105195 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.148795 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="f44dcea2-910c-462b-b3a1-785f93650002" podUID="bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.194551 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.204265 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.486257025 podStartE2EDuration="8.204246414s" podCreationTimestamp="2025-11-28 15:46:36 +0000 UTC" firstStartedPulling="2025-11-28 15:46:37.897666595 +0000 UTC m=+1224.947457916" lastFinishedPulling="2025-11-28 15:46:42.615655994 +0000 UTC m=+1229.665447305" observedRunningTime="2025-11-28 15:46:44.131786322 +0000 UTC m=+1231.181577633" watchObservedRunningTime="2025-11-28 15:46:44.204246414 +0000 UTC m=+1231.254037715" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.249515 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.266235 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-combined-ca-bundle\") pod \"f44dcea2-910c-462b-b3a1-785f93650002\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.266321 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnlnl\" (UniqueName: \"kubernetes.io/projected/f44dcea2-910c-462b-b3a1-785f93650002-kube-api-access-nnlnl\") pod \"f44dcea2-910c-462b-b3a1-785f93650002\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.266412 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config\") pod \"f44dcea2-910c-462b-b3a1-785f93650002\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.266459 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config-secret\") pod \"f44dcea2-910c-462b-b3a1-785f93650002\" (UID: \"f44dcea2-910c-462b-b3a1-785f93650002\") " Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.275966 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f44dcea2-910c-462b-b3a1-785f93650002" (UID: "f44dcea2-910c-462b-b3a1-785f93650002"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.276935 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "f44dcea2-910c-462b-b3a1-785f93650002" (UID: "f44dcea2-910c-462b-b3a1-785f93650002"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.278135 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "f44dcea2-910c-462b-b3a1-785f93650002" (UID: "f44dcea2-910c-462b-b3a1-785f93650002"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.294478 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f44dcea2-910c-462b-b3a1-785f93650002-kube-api-access-nnlnl" (OuterVolumeSpecName: "kube-api-access-nnlnl") pod "f44dcea2-910c-462b-b3a1-785f93650002" (UID: "f44dcea2-910c-462b-b3a1-785f93650002"). InnerVolumeSpecName "kube-api-access-nnlnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.349874 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6d99f7798d-mk9wl"] Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.355413 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6d99f7798d-mk9wl" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api-log" containerID="cri-o://2dd22d1165298bd8fae99be54a7da6131f198cf89ebec17a42db3f5a79bec75c" gracePeriod=30 Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.355503 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6d99f7798d-mk9wl" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api" containerID="cri-o://9c63eb702150b3393258a9c8302ff4ec3b97d4bbbbf27d0d1583e32334575127" gracePeriod=30 Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.369683 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.369722 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.369734 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44dcea2-910c-462b-b3a1-785f93650002-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.369744 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnlnl\" (UniqueName: \"kubernetes.io/projected/f44dcea2-910c-462b-b3a1-785f93650002-kube-api-access-nnlnl\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.439485 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 15:46:44 crc kubenswrapper[4805]: I1128 15:46:44.759616 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.128174 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" event={"ID":"b77f479b-8072-4e56-8517-cce5ec74774c","Type":"ContainerStarted","Data":"1d1558e326764d153e32793a2b7d65350d6db2ba68ebdad5bc0e57ce8100a927"} Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.129251 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.134944 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43372220-e57d-45be-a9ec-ff7cdc0f5e02","Type":"ContainerStarted","Data":"0b5aa5289ebd8cc526f139d9e57efc02de3ec5d736096fb58399311a9fd85fe1"} Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.136331 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc","Type":"ContainerStarted","Data":"dbb597246a7126659257fcdda8d0f182fcad43ff7c2510b74ad1b1edecc2dfe1"} Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.156650 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" podStartSLOduration=4.156626271 podStartE2EDuration="4.156626271s" podCreationTimestamp="2025-11-28 15:46:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:45.147915594 +0000 UTC m=+1232.197706915" watchObservedRunningTime="2025-11-28 15:46:45.156626271 +0000 UTC m=+1232.206417582" Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.157418 4805 generic.go:334] "Generic (PLEG): container finished" podID="51176eca-5fa0-411d-976f-4095550240cd" containerID="2dd22d1165298bd8fae99be54a7da6131f198cf89ebec17a42db3f5a79bec75c" exitCode=143 Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.157491 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d99f7798d-mk9wl" event={"ID":"51176eca-5fa0-411d-976f-4095550240cd","Type":"ContainerDied","Data":"2dd22d1165298bd8fae99be54a7da6131f198cf89ebec17a42db3f5a79bec75c"} Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.163064 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e","Type":"ContainerStarted","Data":"d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53"} Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.163157 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.177070 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="f44dcea2-910c-462b-b3a1-785f93650002" podUID="bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" Nov 28 15:46:45 crc kubenswrapper[4805]: I1128 15:46:45.318395 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f44dcea2-910c-462b-b3a1-785f93650002" path="/var/lib/kubelet/pods/f44dcea2-910c-462b-b3a1-785f93650002/volumes" Nov 28 15:46:46 crc kubenswrapper[4805]: I1128 15:46:46.172857 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43372220-e57d-45be-a9ec-ff7cdc0f5e02","Type":"ContainerStarted","Data":"cbc194d21f9412613534507921bd529d6e8db135f757515054380a5842310e21"} Nov 28 15:46:46 crc kubenswrapper[4805]: I1128 15:46:46.178693 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e","Type":"ContainerStarted","Data":"8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657"} Nov 28 15:46:46 crc kubenswrapper[4805]: I1128 15:46:46.178817 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerName="cinder-api-log" containerID="cri-o://d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53" gracePeriod=30 Nov 28 15:46:46 crc kubenswrapper[4805]: I1128 15:46:46.178875 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerName="cinder-api" containerID="cri-o://8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657" gracePeriod=30 Nov 28 15:46:46 crc kubenswrapper[4805]: I1128 15:46:46.179072 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 28 15:46:46 crc kubenswrapper[4805]: I1128 15:46:46.200016 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.926141642 podStartE2EDuration="5.199993166s" podCreationTimestamp="2025-11-28 15:46:41 +0000 UTC" firstStartedPulling="2025-11-28 15:46:42.223222445 +0000 UTC m=+1229.273013756" lastFinishedPulling="2025-11-28 15:46:43.497073969 +0000 UTC m=+1230.546865280" observedRunningTime="2025-11-28 15:46:46.191505105 +0000 UTC m=+1233.241296406" watchObservedRunningTime="2025-11-28 15:46:46.199993166 +0000 UTC m=+1233.249784497" Nov 28 15:46:46 crc kubenswrapper[4805]: I1128 15:46:46.234486 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.234466465 podStartE2EDuration="5.234466465s" podCreationTimestamp="2025-11-28 15:46:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:46.224289277 +0000 UTC m=+1233.274080588" watchObservedRunningTime="2025-11-28 15:46:46.234466465 +0000 UTC m=+1233.284257776" Nov 28 15:46:46 crc kubenswrapper[4805]: I1128 15:46:46.492466 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 28 15:46:46 crc kubenswrapper[4805]: I1128 15:46:46.927123 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.038442 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data\") pod \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.038521 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-combined-ca-bundle\") pod \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.038601 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-etc-machine-id\") pod \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.038736 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data-custom\") pod \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.038789 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-logs\") pod \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.038824 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvxbd\" (UniqueName: \"kubernetes.io/projected/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-kube-api-access-lvxbd\") pod \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.038850 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-scripts\") pod \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\" (UID: \"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e\") " Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.039465 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" (UID: "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.040147 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-logs" (OuterVolumeSpecName: "logs") pod "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" (UID: "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.046991 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" (UID: "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.050544 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-kube-api-access-lvxbd" (OuterVolumeSpecName: "kube-api-access-lvxbd") pod "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" (UID: "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e"). InnerVolumeSpecName "kube-api-access-lvxbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.066543 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-scripts" (OuterVolumeSpecName: "scripts") pod "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" (UID: "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.092493 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" (UID: "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.117487 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data" (OuterVolumeSpecName: "config-data") pod "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" (UID: "0c6b6fd7-93fb-4c63-bb79-277c17e57f5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.145003 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.145033 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.145044 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvxbd\" (UniqueName: \"kubernetes.io/projected/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-kube-api-access-lvxbd\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.145055 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.145063 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.145071 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.145082 4805 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.207110 4805 generic.go:334] "Generic (PLEG): container finished" podID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerID="8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657" exitCode=0 Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.207150 4805 generic.go:334] "Generic (PLEG): container finished" podID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerID="d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53" exitCode=143 Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.207922 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.218302 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e","Type":"ContainerDied","Data":"8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657"} Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.223500 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e","Type":"ContainerDied","Data":"d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53"} Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.223590 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c6b6fd7-93fb-4c63-bb79-277c17e57f5e","Type":"ContainerDied","Data":"2d30f3d6cf7ca0a747607cee472eb2d1cd713d3a8be1b4b1ae3a98d8921d9fe8"} Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.223665 4805 scope.go:117] "RemoveContainer" containerID="8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.254416 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.286559 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.318614 4805 scope.go:117] "RemoveContainer" containerID="d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.341414 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:46:47 crc kubenswrapper[4805]: E1128 15:46:47.341785 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerName="cinder-api" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.341801 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerName="cinder-api" Nov 28 15:46:47 crc kubenswrapper[4805]: E1128 15:46:47.341828 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerName="cinder-api-log" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.341834 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerName="cinder-api-log" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.342004 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerName="cinder-api-log" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.342031 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" containerName="cinder-api" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.342965 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.347166 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.347427 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.347500 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.347714 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.350014 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf67w\" (UniqueName: \"kubernetes.io/projected/be5d0310-fb90-4d73-94f3-57b105a2a408-kube-api-access-tf67w\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.350063 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.350109 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5d0310-fb90-4d73-94f3-57b105a2a408-logs\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.350139 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-scripts\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.350155 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.350172 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data-custom\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.350339 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.350386 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be5d0310-fb90-4d73-94f3-57b105a2a408-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.350401 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.402653 4805 scope.go:117] "RemoveContainer" containerID="8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657" Nov 28 15:46:47 crc kubenswrapper[4805]: E1128 15:46:47.405153 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657\": container with ID starting with 8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657 not found: ID does not exist" containerID="8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.405193 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657"} err="failed to get container status \"8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657\": rpc error: code = NotFound desc = could not find container \"8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657\": container with ID starting with 8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657 not found: ID does not exist" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.405222 4805 scope.go:117] "RemoveContainer" containerID="d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53" Nov 28 15:46:47 crc kubenswrapper[4805]: E1128 15:46:47.405610 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53\": container with ID starting with d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53 not found: ID does not exist" containerID="d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.405634 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53"} err="failed to get container status \"d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53\": rpc error: code = NotFound desc = could not find container \"d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53\": container with ID starting with d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53 not found: ID does not exist" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.405648 4805 scope.go:117] "RemoveContainer" containerID="8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.413909 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657"} err="failed to get container status \"8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657\": rpc error: code = NotFound desc = could not find container \"8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657\": container with ID starting with 8dcdde8c18ccfdbfde8fb59177fd7822ffc88c992858b1848ac9c00305a9a657 not found: ID does not exist" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.413954 4805 scope.go:117] "RemoveContainer" containerID="d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.414775 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53"} err="failed to get container status \"d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53\": rpc error: code = NotFound desc = could not find container \"d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53\": container with ID starting with d3ed06484f42c15361ac6ad8d28b430cec1b669febf79477ffeb026b4ff50d53 not found: ID does not exist" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.466916 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data-custom\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.467103 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.467145 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be5d0310-fb90-4d73-94f3-57b105a2a408-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.467163 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.467246 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf67w\" (UniqueName: \"kubernetes.io/projected/be5d0310-fb90-4d73-94f3-57b105a2a408-kube-api-access-tf67w\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.467291 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.467328 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5d0310-fb90-4d73-94f3-57b105a2a408-logs\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.467479 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-scripts\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.467516 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.468607 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5d0310-fb90-4d73-94f3-57b105a2a408-logs\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.470427 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be5d0310-fb90-4d73-94f3-57b105a2a408-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.474580 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-scripts\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.476338 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.476676 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data-custom\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.478159 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.483085 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.492946 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf67w\" (UniqueName: \"kubernetes.io/projected/be5d0310-fb90-4d73-94f3-57b105a2a408-kube-api-access-tf67w\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.498859 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.703747 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.830506 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d99f7798d-mk9wl" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:43004->10.217.0.152:9311: read: connection reset by peer" Nov 28 15:46:47 crc kubenswrapper[4805]: I1128 15:46:47.830818 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d99f7798d-mk9wl" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:43006->10.217.0.152:9311: read: connection reset by peer" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.247022 4805 generic.go:334] "Generic (PLEG): container finished" podID="51176eca-5fa0-411d-976f-4095550240cd" containerID="9c63eb702150b3393258a9c8302ff4ec3b97d4bbbbf27d0d1583e32334575127" exitCode=0 Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.247411 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d99f7798d-mk9wl" event={"ID":"51176eca-5fa0-411d-976f-4095550240cd","Type":"ContainerDied","Data":"9c63eb702150b3393258a9c8302ff4ec3b97d4bbbbf27d0d1583e32334575127"} Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.269769 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.388585 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.507855 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data\") pod \"51176eca-5fa0-411d-976f-4095550240cd\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.507938 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51176eca-5fa0-411d-976f-4095550240cd-logs\") pod \"51176eca-5fa0-411d-976f-4095550240cd\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.507999 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jcjj\" (UniqueName: \"kubernetes.io/projected/51176eca-5fa0-411d-976f-4095550240cd-kube-api-access-6jcjj\") pod \"51176eca-5fa0-411d-976f-4095550240cd\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.508094 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data-custom\") pod \"51176eca-5fa0-411d-976f-4095550240cd\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.508136 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-combined-ca-bundle\") pod \"51176eca-5fa0-411d-976f-4095550240cd\" (UID: \"51176eca-5fa0-411d-976f-4095550240cd\") " Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.508573 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51176eca-5fa0-411d-976f-4095550240cd-logs" (OuterVolumeSpecName: "logs") pod "51176eca-5fa0-411d-976f-4095550240cd" (UID: "51176eca-5fa0-411d-976f-4095550240cd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.513904 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51176eca-5fa0-411d-976f-4095550240cd-kube-api-access-6jcjj" (OuterVolumeSpecName: "kube-api-access-6jcjj") pod "51176eca-5fa0-411d-976f-4095550240cd" (UID: "51176eca-5fa0-411d-976f-4095550240cd"). InnerVolumeSpecName "kube-api-access-6jcjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.561319 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "51176eca-5fa0-411d-976f-4095550240cd" (UID: "51176eca-5fa0-411d-976f-4095550240cd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.563584 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51176eca-5fa0-411d-976f-4095550240cd" (UID: "51176eca-5fa0-411d-976f-4095550240cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.585645 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data" (OuterVolumeSpecName: "config-data") pod "51176eca-5fa0-411d-976f-4095550240cd" (UID: "51176eca-5fa0-411d-976f-4095550240cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.610343 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jcjj\" (UniqueName: \"kubernetes.io/projected/51176eca-5fa0-411d-976f-4095550240cd-kube-api-access-6jcjj\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.610411 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.610420 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.610429 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51176eca-5fa0-411d-976f-4095550240cd-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:48 crc kubenswrapper[4805]: I1128 15:46:48.610468 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51176eca-5fa0-411d-976f-4095550240cd-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:49 crc kubenswrapper[4805]: I1128 15:46:49.218441 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c6b6fd7-93fb-4c63-bb79-277c17e57f5e" path="/var/lib/kubelet/pods/0c6b6fd7-93fb-4c63-bb79-277c17e57f5e/volumes" Nov 28 15:46:49 crc kubenswrapper[4805]: I1128 15:46:49.274877 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d99f7798d-mk9wl" event={"ID":"51176eca-5fa0-411d-976f-4095550240cd","Type":"ContainerDied","Data":"6d7348aaafe5568f2fc6168ba682b8272118d91304a13f8b6dd42e7276d0f5e4"} Nov 28 15:46:49 crc kubenswrapper[4805]: I1128 15:46:49.274937 4805 scope.go:117] "RemoveContainer" containerID="9c63eb702150b3393258a9c8302ff4ec3b97d4bbbbf27d0d1583e32334575127" Nov 28 15:46:49 crc kubenswrapper[4805]: I1128 15:46:49.274943 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d99f7798d-mk9wl" Nov 28 15:46:49 crc kubenswrapper[4805]: I1128 15:46:49.277644 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be5d0310-fb90-4d73-94f3-57b105a2a408","Type":"ContainerStarted","Data":"16befe4e2598415b16fc7b244b3cb1d7660ff6773e1cc20677bf867cfd4f1442"} Nov 28 15:46:49 crc kubenswrapper[4805]: I1128 15:46:49.277686 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be5d0310-fb90-4d73-94f3-57b105a2a408","Type":"ContainerStarted","Data":"8f5e0f1d7c83ce9911dbbe2425dbef50ab47911c099ee055d4c347c824de0de7"} Nov 28 15:46:49 crc kubenswrapper[4805]: I1128 15:46:49.309009 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6d99f7798d-mk9wl"] Nov 28 15:46:49 crc kubenswrapper[4805]: I1128 15:46:49.314878 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6d99f7798d-mk9wl"] Nov 28 15:46:49 crc kubenswrapper[4805]: I1128 15:46:49.392527 4805 scope.go:117] "RemoveContainer" containerID="2dd22d1165298bd8fae99be54a7da6131f198cf89ebec17a42db3f5a79bec75c" Nov 28 15:46:50 crc kubenswrapper[4805]: I1128 15:46:50.289893 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be5d0310-fb90-4d73-94f3-57b105a2a408","Type":"ContainerStarted","Data":"0cc217baf0960656d6ad49571b76347e41bfd409a11a61ef61f68ae6d1b9a1c5"} Nov 28 15:46:50 crc kubenswrapper[4805]: I1128 15:46:50.290068 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 28 15:46:50 crc kubenswrapper[4805]: I1128 15:46:50.310937 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.3109137029999998 podStartE2EDuration="3.310913703s" podCreationTimestamp="2025-11-28 15:46:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:50.310162482 +0000 UTC m=+1237.359953793" watchObservedRunningTime="2025-11-28 15:46:50.310913703 +0000 UTC m=+1237.360705044" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.216321 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51176eca-5fa0-411d-976f-4095550240cd" path="/var/lib/kubelet/pods/51176eca-5fa0-411d-976f-4095550240cd/volumes" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.596781 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6f8fc6d4f-g6nbp"] Nov 28 15:46:51 crc kubenswrapper[4805]: E1128 15:46:51.599005 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.599047 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api" Nov 28 15:46:51 crc kubenswrapper[4805]: E1128 15:46:51.599087 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api-log" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.599095 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api-log" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.599623 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api-log" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.599679 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="51176eca-5fa0-411d-976f-4095550240cd" containerName="barbican-api" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.601713 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.619574 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.620306 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.620381 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.625416 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6f8fc6d4f-g6nbp"] Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.655651 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.713104 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-config-data\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.713182 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5h8n\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-kube-api-access-v5h8n\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.713250 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-combined-ca-bundle\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.713305 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-run-httpd\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.713408 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-log-httpd\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.713444 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-etc-swift\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.713490 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-public-tls-certs\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.713541 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-internal-tls-certs\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.717504 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bcffb858c-v9wlq"] Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.717753 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" podUID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerName="dnsmasq-dns" containerID="cri-o://f1236c3a6f401b280bc06c233c0efc5e15b6539effc24bdee4cab358f57dbb1c" gracePeriod=10 Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.821509 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-etc-swift\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.821904 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-public-tls-certs\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.821953 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-internal-tls-certs\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.822032 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-config-data\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.822087 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5h8n\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-kube-api-access-v5h8n\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.822139 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-combined-ca-bundle\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.822200 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-run-httpd\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.822325 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-log-httpd\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.824130 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-run-httpd\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.825196 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-log-httpd\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.836107 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-public-tls-certs\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.846324 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-combined-ca-bundle\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.848293 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-config-data\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.861269 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-internal-tls-certs\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.881375 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5h8n\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-kube-api-access-v5h8n\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.881641 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-etc-swift\") pod \"swift-proxy-6f8fc6d4f-g6nbp\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.924941 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" podUID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.151:5353: connect: connection refused" Nov 28 15:46:51 crc kubenswrapper[4805]: I1128 15:46:51.940366 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:52 crc kubenswrapper[4805]: I1128 15:46:52.037251 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 28 15:46:52 crc kubenswrapper[4805]: I1128 15:46:52.091401 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:46:52 crc kubenswrapper[4805]: I1128 15:46:52.314667 4805 generic.go:334] "Generic (PLEG): container finished" podID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerID="f1236c3a6f401b280bc06c233c0efc5e15b6539effc24bdee4cab358f57dbb1c" exitCode=0 Nov 28 15:46:52 crc kubenswrapper[4805]: I1128 15:46:52.315046 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerName="cinder-scheduler" containerID="cri-o://0b5aa5289ebd8cc526f139d9e57efc02de3ec5d736096fb58399311a9fd85fe1" gracePeriod=30 Nov 28 15:46:52 crc kubenswrapper[4805]: I1128 15:46:52.315110 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" event={"ID":"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950","Type":"ContainerDied","Data":"f1236c3a6f401b280bc06c233c0efc5e15b6539effc24bdee4cab358f57dbb1c"} Nov 28 15:46:52 crc kubenswrapper[4805]: I1128 15:46:52.315255 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerName="probe" containerID="cri-o://cbc194d21f9412613534507921bd529d6e8db135f757515054380a5842310e21" gracePeriod=30 Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.016266 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.016571 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="ceilometer-central-agent" containerID="cri-o://272aa407188898c5f5222aa4d11cbe180785d3c06917d009a59e004294569021" gracePeriod=30 Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.017519 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="proxy-httpd" containerID="cri-o://b842b88b2bff9eb8270ea244a59e110d63b3fff409cac01ed749dc5086362891" gracePeriod=30 Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.017525 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="sg-core" containerID="cri-o://62be3d9ee2c288accb4e7e1c932b1555386148b92540d70e2a4b9211551ee385" gracePeriod=30 Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.017528 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="ceilometer-notification-agent" containerID="cri-o://cbd1edd2530ac67b1f84c5e72a6cca109f3a42bf326efdf16622f58cb57233e3" gracePeriod=30 Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.028727 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.157:3000/\": EOF" Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.327071 4805 generic.go:334] "Generic (PLEG): container finished" podID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerID="b842b88b2bff9eb8270ea244a59e110d63b3fff409cac01ed749dc5086362891" exitCode=0 Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.327103 4805 generic.go:334] "Generic (PLEG): container finished" podID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerID="62be3d9ee2c288accb4e7e1c932b1555386148b92540d70e2a4b9211551ee385" exitCode=2 Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.327157 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerDied","Data":"b842b88b2bff9eb8270ea244a59e110d63b3fff409cac01ed749dc5086362891"} Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.327210 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerDied","Data":"62be3d9ee2c288accb4e7e1c932b1555386148b92540d70e2a4b9211551ee385"} Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.331082 4805 generic.go:334] "Generic (PLEG): container finished" podID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerID="cbc194d21f9412613534507921bd529d6e8db135f757515054380a5842310e21" exitCode=0 Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.331120 4805 generic.go:334] "Generic (PLEG): container finished" podID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerID="0b5aa5289ebd8cc526f139d9e57efc02de3ec5d736096fb58399311a9fd85fe1" exitCode=0 Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.331146 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43372220-e57d-45be-a9ec-ff7cdc0f5e02","Type":"ContainerDied","Data":"cbc194d21f9412613534507921bd529d6e8db135f757515054380a5842310e21"} Nov 28 15:46:53 crc kubenswrapper[4805]: I1128 15:46:53.331178 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43372220-e57d-45be-a9ec-ff7cdc0f5e02","Type":"ContainerDied","Data":"0b5aa5289ebd8cc526f139d9e57efc02de3ec5d736096fb58399311a9fd85fe1"} Nov 28 15:46:54 crc kubenswrapper[4805]: I1128 15:46:54.348454 4805 generic.go:334] "Generic (PLEG): container finished" podID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerID="272aa407188898c5f5222aa4d11cbe180785d3c06917d009a59e004294569021" exitCode=0 Nov 28 15:46:54 crc kubenswrapper[4805]: I1128 15:46:54.348512 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerDied","Data":"272aa407188898c5f5222aa4d11cbe180785d3c06917d009a59e004294569021"} Nov 28 15:46:55 crc kubenswrapper[4805]: I1128 15:46:55.364237 4805 generic.go:334] "Generic (PLEG): container finished" podID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerID="cbd1edd2530ac67b1f84c5e72a6cca109f3a42bf326efdf16622f58cb57233e3" exitCode=0 Nov 28 15:46:55 crc kubenswrapper[4805]: I1128 15:46:55.364319 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerDied","Data":"cbd1edd2530ac67b1f84c5e72a6cca109f3a42bf326efdf16622f58cb57233e3"} Nov 28 15:46:56 crc kubenswrapper[4805]: I1128 15:46:56.882320 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:46:56 crc kubenswrapper[4805]: I1128 15:46:56.882597 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-log" containerID="cri-o://c655e1242a0c7987b02a1c84687637804a2b82ea438583fed89c15a2e0f3dc56" gracePeriod=30 Nov 28 15:46:56 crc kubenswrapper[4805]: I1128 15:46:56.882700 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-httpd" containerID="cri-o://1af8fb2da9461341a610764bda29e02ea05cb8bafd6f608ea0e1266702bb14e5" gracePeriod=30 Nov 28 15:46:56 crc kubenswrapper[4805]: I1128 15:46:56.922964 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" podUID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.151:5353: connect: connection refused" Nov 28 15:46:57 crc kubenswrapper[4805]: I1128 15:46:57.384278 4805 generic.go:334] "Generic (PLEG): container finished" podID="b9d91b74-236c-4989-93ca-2379328e7119" containerID="c655e1242a0c7987b02a1c84687637804a2b82ea438583fed89c15a2e0f3dc56" exitCode=143 Nov 28 15:46:57 crc kubenswrapper[4805]: I1128 15:46:57.384371 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b9d91b74-236c-4989-93ca-2379328e7119","Type":"ContainerDied","Data":"c655e1242a0c7987b02a1c84687637804a2b82ea438583fed89c15a2e0f3dc56"} Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.028462 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.153867 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.158979 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-nb\") pod \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.159062 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l47rq\" (UniqueName: \"kubernetes.io/projected/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-kube-api-access-l47rq\") pod \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.159106 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-config\") pod \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.159264 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-sb\") pod \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.159299 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-swift-storage-0\") pod \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.159319 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-svc\") pod \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\" (UID: \"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.168592 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-kube-api-access-l47rq" (OuterVolumeSpecName: "kube-api-access-l47rq") pod "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" (UID: "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950"). InnerVolumeSpecName "kube-api-access-l47rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.215030 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.261929 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data-custom\") pod \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.261985 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8zs5\" (UniqueName: \"kubernetes.io/projected/43372220-e57d-45be-a9ec-ff7cdc0f5e02-kube-api-access-d8zs5\") pod \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262081 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43372220-e57d-45be-a9ec-ff7cdc0f5e02-etc-machine-id\") pod \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262129 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-sg-core-conf-yaml\") pod \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262155 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-log-httpd\") pod \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262182 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data\") pod \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262301 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-config-data\") pod \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262390 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-scripts\") pod \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262418 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-scripts\") pod \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262470 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-combined-ca-bundle\") pod \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\" (UID: \"43372220-e57d-45be-a9ec-ff7cdc0f5e02\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262509 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-combined-ca-bundle\") pod \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262534 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5px\" (UniqueName: \"kubernetes.io/projected/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-kube-api-access-qg5px\") pod \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262557 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-run-httpd\") pod \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\" (UID: \"b00e3690-9b01-45a2-9309-c4ee24dc0d6e\") " Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.262651 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43372220-e57d-45be-a9ec-ff7cdc0f5e02-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "43372220-e57d-45be-a9ec-ff7cdc0f5e02" (UID: "43372220-e57d-45be-a9ec-ff7cdc0f5e02"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.263024 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b00e3690-9b01-45a2-9309-c4ee24dc0d6e" (UID: "b00e3690-9b01-45a2-9309-c4ee24dc0d6e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.263261 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l47rq\" (UniqueName: \"kubernetes.io/projected/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-kube-api-access-l47rq\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.263276 4805 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43372220-e57d-45be-a9ec-ff7cdc0f5e02-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.263286 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.266619 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b00e3690-9b01-45a2-9309-c4ee24dc0d6e" (UID: "b00e3690-9b01-45a2-9309-c4ee24dc0d6e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.288287 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-scripts" (OuterVolumeSpecName: "scripts") pod "43372220-e57d-45be-a9ec-ff7cdc0f5e02" (UID: "43372220-e57d-45be-a9ec-ff7cdc0f5e02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.291698 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43372220-e57d-45be-a9ec-ff7cdc0f5e02-kube-api-access-d8zs5" (OuterVolumeSpecName: "kube-api-access-d8zs5") pod "43372220-e57d-45be-a9ec-ff7cdc0f5e02" (UID: "43372220-e57d-45be-a9ec-ff7cdc0f5e02"). InnerVolumeSpecName "kube-api-access-d8zs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.293539 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "43372220-e57d-45be-a9ec-ff7cdc0f5e02" (UID: "43372220-e57d-45be-a9ec-ff7cdc0f5e02"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.294027 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-scripts" (OuterVolumeSpecName: "scripts") pod "b00e3690-9b01-45a2-9309-c4ee24dc0d6e" (UID: "b00e3690-9b01-45a2-9309-c4ee24dc0d6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.294198 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-kube-api-access-qg5px" (OuterVolumeSpecName: "kube-api-access-qg5px") pod "b00e3690-9b01-45a2-9309-c4ee24dc0d6e" (UID: "b00e3690-9b01-45a2-9309-c4ee24dc0d6e"). InnerVolumeSpecName "kube-api-access-qg5px". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.367210 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.367239 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.367249 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5px\" (UniqueName: \"kubernetes.io/projected/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-kube-api-access-qg5px\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.367258 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.367266 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.367276 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8zs5\" (UniqueName: \"kubernetes.io/projected/43372220-e57d-45be-a9ec-ff7cdc0f5e02-kube-api-access-d8zs5\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.402986 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" event={"ID":"dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950","Type":"ContainerDied","Data":"4f513c6ed66741f0cd62c728193c47e57852b094a4b69d6b76193ac63582f460"} Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.403297 4805 scope.go:117] "RemoveContainer" containerID="f1236c3a6f401b280bc06c233c0efc5e15b6539effc24bdee4cab358f57dbb1c" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.403638 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bcffb858c-v9wlq" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.409819 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43372220-e57d-45be-a9ec-ff7cdc0f5e02","Type":"ContainerDied","Data":"ed49a51280298e3afa79ad05b7d940e4933a67c330bb4f37172ee83f9747dc5a"} Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.409938 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.425634 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b00e3690-9b01-45a2-9309-c4ee24dc0d6e" (UID: "b00e3690-9b01-45a2-9309-c4ee24dc0d6e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.431546 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc","Type":"ContainerStarted","Data":"051cad31407ced2cb3d554ef9c21f2839c03af141e5fd2cda0c2a783a4cd32a3"} Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.432674 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" (UID: "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.438101 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b00e3690-9b01-45a2-9309-c4ee24dc0d6e","Type":"ContainerDied","Data":"5ab4c0d7a7df62739992e348d107282bddc8eb0fe733dd64e5d70ed878ff4494"} Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.438219 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.452569 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" (UID: "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.453246 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.258758047 podStartE2EDuration="15.45323116s" podCreationTimestamp="2025-11-28 15:46:43 +0000 UTC" firstStartedPulling="2025-11-28 15:46:44.43376115 +0000 UTC m=+1231.483552461" lastFinishedPulling="2025-11-28 15:46:57.628234263 +0000 UTC m=+1244.678025574" observedRunningTime="2025-11-28 15:46:58.44696576 +0000 UTC m=+1245.496757071" watchObservedRunningTime="2025-11-28 15:46:58.45323116 +0000 UTC m=+1245.503022471" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.469014 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.469042 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.469052 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: W1128 15:46:58.481533 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33489fb0_b55a_4544_8a36_c4e0c9ca10f0.slice/crio-b5b41c3e7b2867ed3df58372ba9fa27ee74b0e5c928fe954c5aa4b3080b78fb5 WatchSource:0}: Error finding container b5b41c3e7b2867ed3df58372ba9fa27ee74b0e5c928fe954c5aa4b3080b78fb5: Status 404 returned error can't find the container with id b5b41c3e7b2867ed3df58372ba9fa27ee74b0e5c928fe954c5aa4b3080b78fb5 Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.482745 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" (UID: "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.483397 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6f8fc6d4f-g6nbp"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.485112 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" (UID: "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.490951 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-config" (OuterVolumeSpecName: "config") pod "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" (UID: "dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.507249 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43372220-e57d-45be-a9ec-ff7cdc0f5e02" (UID: "43372220-e57d-45be-a9ec-ff7cdc0f5e02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.534338 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data" (OuterVolumeSpecName: "config-data") pod "43372220-e57d-45be-a9ec-ff7cdc0f5e02" (UID: "43372220-e57d-45be-a9ec-ff7cdc0f5e02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.555590 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-config-data" (OuterVolumeSpecName: "config-data") pod "b00e3690-9b01-45a2-9309-c4ee24dc0d6e" (UID: "b00e3690-9b01-45a2-9309-c4ee24dc0d6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.570932 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.570958 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.570968 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.570976 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.570985 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.570993 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43372220-e57d-45be-a9ec-ff7cdc0f5e02-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.572588 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b00e3690-9b01-45a2-9309-c4ee24dc0d6e" (UID: "b00e3690-9b01-45a2-9309-c4ee24dc0d6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.653746 4805 scope.go:117] "RemoveContainer" containerID="18e7dc78edf79d28cdfed7f21eb157b3e47abb87593b65dc74d696841dc2a324" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.673412 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00e3690-9b01-45a2-9309-c4ee24dc0d6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.682431 4805 scope.go:117] "RemoveContainer" containerID="cbc194d21f9412613534507921bd529d6e8db135f757515054380a5842310e21" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.731340 4805 scope.go:117] "RemoveContainer" containerID="0b5aa5289ebd8cc526f139d9e57efc02de3ec5d736096fb58399311a9fd85fe1" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.758598 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bcffb858c-v9wlq"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.762253 4805 scope.go:117] "RemoveContainer" containerID="b842b88b2bff9eb8270ea244a59e110d63b3fff409cac01ed749dc5086362891" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.774092 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bcffb858c-v9wlq"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.788254 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.788819 4805 scope.go:117] "RemoveContainer" containerID="62be3d9ee2c288accb4e7e1c932b1555386148b92540d70e2a4b9211551ee385" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.797689 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.807246 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.814257 4805 scope.go:117] "RemoveContainer" containerID="cbd1edd2530ac67b1f84c5e72a6cca109f3a42bf326efdf16622f58cb57233e3" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.816696 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:46:58 crc kubenswrapper[4805]: E1128 15:46:58.817140 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerName="probe" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.817161 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerName="probe" Nov 28 15:46:58 crc kubenswrapper[4805]: E1128 15:46:58.817173 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="proxy-httpd" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.817182 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="proxy-httpd" Nov 28 15:46:58 crc kubenswrapper[4805]: E1128 15:46:58.817194 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="ceilometer-notification-agent" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.817201 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="ceilometer-notification-agent" Nov 28 15:46:58 crc kubenswrapper[4805]: E1128 15:46:58.817213 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="ceilometer-central-agent" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.817220 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="ceilometer-central-agent" Nov 28 15:46:58 crc kubenswrapper[4805]: E1128 15:46:58.817230 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerName="dnsmasq-dns" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.817236 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerName="dnsmasq-dns" Nov 28 15:46:58 crc kubenswrapper[4805]: E1128 15:46:58.817255 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="sg-core" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.817262 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="sg-core" Nov 28 15:46:58 crc kubenswrapper[4805]: E1128 15:46:58.817272 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerName="cinder-scheduler" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.817278 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerName="cinder-scheduler" Nov 28 15:46:58 crc kubenswrapper[4805]: E1128 15:46:58.817295 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerName="init" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.817301 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerName="init" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.818206 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="proxy-httpd" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.818239 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerName="cinder-scheduler" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.818264 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" containerName="probe" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.818277 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="ceilometer-central-agent" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.818284 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="sg-core" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.818296 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" containerName="dnsmasq-dns" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.818306 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" containerName="ceilometer-notification-agent" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.819273 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.822278 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.822975 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.834975 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.852014 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.871729 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.873503 4805 scope.go:117] "RemoveContainer" containerID="272aa407188898c5f5222aa4d11cbe180785d3c06917d009a59e004294569021" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.873623 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.873891 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.877543 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54ff5706-18a9-4e71-8289-e3d9e5d9af23-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.877674 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w8rd\" (UniqueName: \"kubernetes.io/projected/54ff5706-18a9-4e71-8289-e3d9e5d9af23-kube-api-access-7w8rd\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.877755 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-scripts\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.877801 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.877878 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.877927 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.914317 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.979835 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-log-httpd\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.979908 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.979945 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.979998 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54ff5706-18a9-4e71-8289-e3d9e5d9af23-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.980106 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54ff5706-18a9-4e71-8289-e3d9e5d9af23-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.980034 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.980184 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-scripts\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.980440 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-run-httpd\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.980912 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbfp2\" (UniqueName: \"kubernetes.io/projected/54782d85-dc74-4e4f-b8ce-4cb181529e22-kube-api-access-hbfp2\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.981002 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w8rd\" (UniqueName: \"kubernetes.io/projected/54ff5706-18a9-4e71-8289-e3d9e5d9af23-kube-api-access-7w8rd\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.981402 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-scripts\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.981969 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.982053 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.982132 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-config-data\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.983602 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.985407 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-scripts\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.987854 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:58 crc kubenswrapper[4805]: I1128 15:46:58.989112 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.003125 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w8rd\" (UniqueName: \"kubernetes.io/projected/54ff5706-18a9-4e71-8289-e3d9e5d9af23-kube-api-access-7w8rd\") pod \"cinder-scheduler-0\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " pod="openstack/cinder-scheduler-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.022760 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:46:59 crc kubenswrapper[4805]: E1128 15:46:59.023431 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-hbfp2 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="54782d85-dc74-4e4f-b8ce-4cb181529e22" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.084033 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.084077 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-config-data\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.084112 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-log-httpd\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.084191 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.084212 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-scripts\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.084225 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-run-httpd\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.084268 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbfp2\" (UniqueName: \"kubernetes.io/projected/54782d85-dc74-4e4f-b8ce-4cb181529e22-kube-api-access-hbfp2\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.085069 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-log-httpd\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.085237 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-run-httpd\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.088268 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-scripts\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.089400 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.090433 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-config-data\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.090813 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.103987 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbfp2\" (UniqueName: \"kubernetes.io/projected/54782d85-dc74-4e4f-b8ce-4cb181529e22-kube-api-access-hbfp2\") pod \"ceilometer-0\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.147577 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.214655 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43372220-e57d-45be-a9ec-ff7cdc0f5e02" path="/var/lib/kubelet/pods/43372220-e57d-45be-a9ec-ff7cdc0f5e02/volumes" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.217011 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b00e3690-9b01-45a2-9309-c4ee24dc0d6e" path="/var/lib/kubelet/pods/b00e3690-9b01-45a2-9309-c4ee24dc0d6e/volumes" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.217906 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950" path="/var/lib/kubelet/pods/dad7cd0b-6f5c-43a8-a0c1-5d33cc57c950/volumes" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.393669 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-rr5bn"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.400186 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.417973 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rr5bn"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.494967 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" event={"ID":"33489fb0-b55a-4544-8a36-c4e0c9ca10f0","Type":"ContainerStarted","Data":"a02a9bd5b077fb94a3c91b016fbc51400e43005c07086b7606cdb1ad65200db2"} Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.495014 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" event={"ID":"33489fb0-b55a-4544-8a36-c4e0c9ca10f0","Type":"ContainerStarted","Data":"f341ac5588a30f34bf1c5785f352e9e8bcbb991ecae2e3a87e2814de74b6c3c7"} Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.495031 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" event={"ID":"33489fb0-b55a-4544-8a36-c4e0c9ca10f0","Type":"ContainerStarted","Data":"b5b41c3e7b2867ed3df58372ba9fa27ee74b0e5c928fe954c5aa4b3080b78fb5"} Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.496289 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.496319 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.496424 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzbb6\" (UniqueName: \"kubernetes.io/projected/3672f82d-1221-47a2-8be9-cbc3394def33-kube-api-access-qzbb6\") pod \"nova-api-db-create-rr5bn\" (UID: \"3672f82d-1221-47a2-8be9-cbc3394def33\") " pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.496594 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3672f82d-1221-47a2-8be9-cbc3394def33-operator-scripts\") pod \"nova-api-db-create-rr5bn\" (UID: \"3672f82d-1221-47a2-8be9-cbc3394def33\") " pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.498171 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.522308 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.535695 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" podStartSLOduration=8.535678387 podStartE2EDuration="8.535678387s" podCreationTimestamp="2025-11-28 15:46:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:46:59.534809362 +0000 UTC m=+1246.584600673" watchObservedRunningTime="2025-11-28 15:46:59.535678387 +0000 UTC m=+1246.585469698" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.597202 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-vkkgt"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.597543 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-scripts\") pod \"54782d85-dc74-4e4f-b8ce-4cb181529e22\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.597675 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-log-httpd\") pod \"54782d85-dc74-4e4f-b8ce-4cb181529e22\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.597740 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbfp2\" (UniqueName: \"kubernetes.io/projected/54782d85-dc74-4e4f-b8ce-4cb181529e22-kube-api-access-hbfp2\") pod \"54782d85-dc74-4e4f-b8ce-4cb181529e22\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.597769 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-run-httpd\") pod \"54782d85-dc74-4e4f-b8ce-4cb181529e22\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.597808 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-config-data\") pod \"54782d85-dc74-4e4f-b8ce-4cb181529e22\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.597860 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-sg-core-conf-yaml\") pod \"54782d85-dc74-4e4f-b8ce-4cb181529e22\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.597888 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-combined-ca-bundle\") pod \"54782d85-dc74-4e4f-b8ce-4cb181529e22\" (UID: \"54782d85-dc74-4e4f-b8ce-4cb181529e22\") " Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.598249 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3672f82d-1221-47a2-8be9-cbc3394def33-operator-scripts\") pod \"nova-api-db-create-rr5bn\" (UID: \"3672f82d-1221-47a2-8be9-cbc3394def33\") " pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.598317 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzbb6\" (UniqueName: \"kubernetes.io/projected/3672f82d-1221-47a2-8be9-cbc3394def33-kube-api-access-qzbb6\") pod \"nova-api-db-create-rr5bn\" (UID: \"3672f82d-1221-47a2-8be9-cbc3394def33\") " pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.599615 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.600489 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "54782d85-dc74-4e4f-b8ce-4cb181529e22" (UID: "54782d85-dc74-4e4f-b8ce-4cb181529e22"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.604971 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "54782d85-dc74-4e4f-b8ce-4cb181529e22" (UID: "54782d85-dc74-4e4f-b8ce-4cb181529e22"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.607590 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54782d85-dc74-4e4f-b8ce-4cb181529e22-kube-api-access-hbfp2" (OuterVolumeSpecName: "kube-api-access-hbfp2") pod "54782d85-dc74-4e4f-b8ce-4cb181529e22" (UID: "54782d85-dc74-4e4f-b8ce-4cb181529e22"). InnerVolumeSpecName "kube-api-access-hbfp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.613465 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-config-data" (OuterVolumeSpecName: "config-data") pod "54782d85-dc74-4e4f-b8ce-4cb181529e22" (UID: "54782d85-dc74-4e4f-b8ce-4cb181529e22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.613833 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54782d85-dc74-4e4f-b8ce-4cb181529e22" (UID: "54782d85-dc74-4e4f-b8ce-4cb181529e22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.613954 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3672f82d-1221-47a2-8be9-cbc3394def33-operator-scripts\") pod \"nova-api-db-create-rr5bn\" (UID: \"3672f82d-1221-47a2-8be9-cbc3394def33\") " pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.614544 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "54782d85-dc74-4e4f-b8ce-4cb181529e22" (UID: "54782d85-dc74-4e4f-b8ce-4cb181529e22"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.617416 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vkkgt"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.617453 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-scripts" (OuterVolumeSpecName: "scripts") pod "54782d85-dc74-4e4f-b8ce-4cb181529e22" (UID: "54782d85-dc74-4e4f-b8ce-4cb181529e22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.624296 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzbb6\" (UniqueName: \"kubernetes.io/projected/3672f82d-1221-47a2-8be9-cbc3394def33-kube-api-access-qzbb6\") pod \"nova-api-db-create-rr5bn\" (UID: \"3672f82d-1221-47a2-8be9-cbc3394def33\") " pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.652932 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c2b1-account-create-update-qgsb8"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.654430 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.656342 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.678127 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c2b1-account-create-update-qgsb8"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.715109 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1232c9a3-93bb-4956-8e76-89829ee8b299-operator-scripts\") pod \"nova-cell0-db-create-vkkgt\" (UID: \"1232c9a3-93bb-4956-8e76-89829ee8b299\") " pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.715316 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w424v\" (UniqueName: \"kubernetes.io/projected/1232c9a3-93bb-4956-8e76-89829ee8b299-kube-api-access-w424v\") pod \"nova-cell0-db-create-vkkgt\" (UID: \"1232c9a3-93bb-4956-8e76-89829ee8b299\") " pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.716981 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.717034 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbfp2\" (UniqueName: \"kubernetes.io/projected/54782d85-dc74-4e4f-b8ce-4cb181529e22-kube-api-access-hbfp2\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.717046 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54782d85-dc74-4e4f-b8ce-4cb181529e22-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.717060 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.717069 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.717077 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.717086 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54782d85-dc74-4e4f-b8ce-4cb181529e22-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.754902 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.796960 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.821884 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w424v\" (UniqueName: \"kubernetes.io/projected/1232c9a3-93bb-4956-8e76-89829ee8b299-kube-api-access-w424v\") pod \"nova-cell0-db-create-vkkgt\" (UID: \"1232c9a3-93bb-4956-8e76-89829ee8b299\") " pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.832590 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6cb1991-b179-4169-be04-b0e707cc99a3-operator-scripts\") pod \"nova-api-c2b1-account-create-update-qgsb8\" (UID: \"d6cb1991-b179-4169-be04-b0e707cc99a3\") " pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.832904 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6njc8\" (UniqueName: \"kubernetes.io/projected/d6cb1991-b179-4169-be04-b0e707cc99a3-kube-api-access-6njc8\") pod \"nova-api-c2b1-account-create-update-qgsb8\" (UID: \"d6cb1991-b179-4169-be04-b0e707cc99a3\") " pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.834535 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1232c9a3-93bb-4956-8e76-89829ee8b299-operator-scripts\") pod \"nova-cell0-db-create-vkkgt\" (UID: \"1232c9a3-93bb-4956-8e76-89829ee8b299\") " pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.844270 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1232c9a3-93bb-4956-8e76-89829ee8b299-operator-scripts\") pod \"nova-cell0-db-create-vkkgt\" (UID: \"1232c9a3-93bb-4956-8e76-89829ee8b299\") " pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.854049 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w424v\" (UniqueName: \"kubernetes.io/projected/1232c9a3-93bb-4956-8e76-89829ee8b299-kube-api-access-w424v\") pod \"nova-cell0-db-create-vkkgt\" (UID: \"1232c9a3-93bb-4956-8e76-89829ee8b299\") " pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.881011 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-wn5s2"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.885235 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.889598 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wn5s2"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.907194 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-8782-account-create-update-p6l42"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.910216 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.912504 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.916833 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8782-account-create-update-p6l42"] Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.945010 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6cb1991-b179-4169-be04-b0e707cc99a3-operator-scripts\") pod \"nova-api-c2b1-account-create-update-qgsb8\" (UID: \"d6cb1991-b179-4169-be04-b0e707cc99a3\") " pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.945078 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6njc8\" (UniqueName: \"kubernetes.io/projected/d6cb1991-b179-4169-be04-b0e707cc99a3-kube-api-access-6njc8\") pod \"nova-api-c2b1-account-create-update-qgsb8\" (UID: \"d6cb1991-b179-4169-be04-b0e707cc99a3\") " pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.946704 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6cb1991-b179-4169-be04-b0e707cc99a3-operator-scripts\") pod \"nova-api-c2b1-account-create-update-qgsb8\" (UID: \"d6cb1991-b179-4169-be04-b0e707cc99a3\") " pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:46:59 crc kubenswrapper[4805]: I1128 15:46:59.964732 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6njc8\" (UniqueName: \"kubernetes.io/projected/d6cb1991-b179-4169-be04-b0e707cc99a3-kube-api-access-6njc8\") pod \"nova-api-c2b1-account-create-update-qgsb8\" (UID: \"d6cb1991-b179-4169-be04-b0e707cc99a3\") " pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.017575 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-1502-account-create-update-2sqhk"] Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.021339 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.025258 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.048175 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcpp6\" (UniqueName: \"kubernetes.io/projected/2efa09bc-8af5-473d-a893-0cdd3936a6ce-kube-api-access-lcpp6\") pod \"nova-cell0-8782-account-create-update-p6l42\" (UID: \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\") " pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.048276 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nq6j\" (UniqueName: \"kubernetes.io/projected/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-kube-api-access-9nq6j\") pod \"nova-cell1-db-create-wn5s2\" (UID: \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\") " pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.048302 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-operator-scripts\") pod \"nova-cell1-db-create-wn5s2\" (UID: \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\") " pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.048318 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efa09bc-8af5-473d-a893-0cdd3936a6ce-operator-scripts\") pod \"nova-cell0-8782-account-create-update-p6l42\" (UID: \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\") " pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.056271 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1502-account-create-update-2sqhk"] Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.059716 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.145:9292/healthcheck\": read tcp 10.217.0.2:36986->10.217.0.145:9292: read: connection reset by peer" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.060048 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.145:9292/healthcheck\": read tcp 10.217.0.2:37002->10.217.0.145:9292: read: connection reset by peer" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.145410 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.150195 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nq6j\" (UniqueName: \"kubernetes.io/projected/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-kube-api-access-9nq6j\") pod \"nova-cell1-db-create-wn5s2\" (UID: \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\") " pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.150250 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-operator-scripts\") pod \"nova-cell1-db-create-wn5s2\" (UID: \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\") " pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.150270 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efa09bc-8af5-473d-a893-0cdd3936a6ce-operator-scripts\") pod \"nova-cell0-8782-account-create-update-p6l42\" (UID: \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\") " pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.150300 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79634ed6-4d3f-4488-8123-bce2cb65249d-operator-scripts\") pod \"nova-cell1-1502-account-create-update-2sqhk\" (UID: \"79634ed6-4d3f-4488-8123-bce2cb65249d\") " pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.150386 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcpp6\" (UniqueName: \"kubernetes.io/projected/2efa09bc-8af5-473d-a893-0cdd3936a6ce-kube-api-access-lcpp6\") pod \"nova-cell0-8782-account-create-update-p6l42\" (UID: \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\") " pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.150438 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2rl2\" (UniqueName: \"kubernetes.io/projected/79634ed6-4d3f-4488-8123-bce2cb65249d-kube-api-access-w2rl2\") pod \"nova-cell1-1502-account-create-update-2sqhk\" (UID: \"79634ed6-4d3f-4488-8123-bce2cb65249d\") " pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.151388 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-operator-scripts\") pod \"nova-cell1-db-create-wn5s2\" (UID: \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\") " pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.151890 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efa09bc-8af5-473d-a893-0cdd3936a6ce-operator-scripts\") pod \"nova-cell0-8782-account-create-update-p6l42\" (UID: \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\") " pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.172474 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nq6j\" (UniqueName: \"kubernetes.io/projected/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-kube-api-access-9nq6j\") pod \"nova-cell1-db-create-wn5s2\" (UID: \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\") " pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.174316 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.178238 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcpp6\" (UniqueName: \"kubernetes.io/projected/2efa09bc-8af5-473d-a893-0cdd3936a6ce-kube-api-access-lcpp6\") pod \"nova-cell0-8782-account-create-update-p6l42\" (UID: \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\") " pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.217921 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.240348 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.259306 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2rl2\" (UniqueName: \"kubernetes.io/projected/79634ed6-4d3f-4488-8123-bce2cb65249d-kube-api-access-w2rl2\") pod \"nova-cell1-1502-account-create-update-2sqhk\" (UID: \"79634ed6-4d3f-4488-8123-bce2cb65249d\") " pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.259416 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79634ed6-4d3f-4488-8123-bce2cb65249d-operator-scripts\") pod \"nova-cell1-1502-account-create-update-2sqhk\" (UID: \"79634ed6-4d3f-4488-8123-bce2cb65249d\") " pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.260073 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79634ed6-4d3f-4488-8123-bce2cb65249d-operator-scripts\") pod \"nova-cell1-1502-account-create-update-2sqhk\" (UID: \"79634ed6-4d3f-4488-8123-bce2cb65249d\") " pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.293167 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2rl2\" (UniqueName: \"kubernetes.io/projected/79634ed6-4d3f-4488-8123-bce2cb65249d-kube-api-access-w2rl2\") pod \"nova-cell1-1502-account-create-update-2sqhk\" (UID: \"79634ed6-4d3f-4488-8123-bce2cb65249d\") " pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.365545 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.437631 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rr5bn"] Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.567576 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54ff5706-18a9-4e71-8289-e3d9e5d9af23","Type":"ContainerStarted","Data":"7f867b3d73f37c5c0b2aaf08a77575ff2b6f0f5a9243efe6cda11c61e722f228"} Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.621783 4805 generic.go:334] "Generic (PLEG): container finished" podID="b9d91b74-236c-4989-93ca-2379328e7119" containerID="1af8fb2da9461341a610764bda29e02ea05cb8bafd6f608ea0e1266702bb14e5" exitCode=0 Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.622435 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b9d91b74-236c-4989-93ca-2379328e7119","Type":"ContainerDied","Data":"1af8fb2da9461341a610764bda29e02ea05cb8bafd6f608ea0e1266702bb14e5"} Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.622617 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.688573 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.818722 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.852099 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.852165 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.854504 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.857857 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.858077 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.887210 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-config-data\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.887334 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-run-httpd\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.887381 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.887461 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2szs\" (UniqueName: \"kubernetes.io/projected/7ade9363-30e9-45e1-a57b-00704d8277ca-kube-api-access-l2szs\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.887541 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.887573 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-log-httpd\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.887656 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-scripts\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.892968 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.989252 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-run-httpd\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.989596 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.989673 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2szs\" (UniqueName: \"kubernetes.io/projected/7ade9363-30e9-45e1-a57b-00704d8277ca-kube-api-access-l2szs\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.989722 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.989745 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-log-httpd\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.989769 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-run-httpd\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.989827 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-scripts\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.990100 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-config-data\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.991427 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-log-httpd\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:00 crc kubenswrapper[4805]: I1128 15:47:00.997411 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-scripts\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:00.998489 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:00.999042 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.002277 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-config-data\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.021611 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2szs\" (UniqueName: \"kubernetes.io/projected/7ade9363-30e9-45e1-a57b-00704d8277ca-kube-api-access-l2szs\") pod \"ceilometer-0\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " pod="openstack/ceilometer-0" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.203552 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.232101 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54782d85-dc74-4e4f-b8ce-4cb181529e22" path="/var/lib/kubelet/pods/54782d85-dc74-4e4f-b8ce-4cb181529e22/volumes" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.390499 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.496784 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vkkgt"] Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.503148 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8782-account-create-update-p6l42"] Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.506007 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-logs\") pod \"b9d91b74-236c-4989-93ca-2379328e7119\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.506074 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-scripts\") pod \"b9d91b74-236c-4989-93ca-2379328e7119\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.506113 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75lzn\" (UniqueName: \"kubernetes.io/projected/b9d91b74-236c-4989-93ca-2379328e7119-kube-api-access-75lzn\") pod \"b9d91b74-236c-4989-93ca-2379328e7119\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.506184 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-internal-tls-certs\") pod \"b9d91b74-236c-4989-93ca-2379328e7119\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.507312 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-config-data\") pod \"b9d91b74-236c-4989-93ca-2379328e7119\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.507368 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-httpd-run\") pod \"b9d91b74-236c-4989-93ca-2379328e7119\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.507386 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"b9d91b74-236c-4989-93ca-2379328e7119\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.507452 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-combined-ca-bundle\") pod \"b9d91b74-236c-4989-93ca-2379328e7119\" (UID: \"b9d91b74-236c-4989-93ca-2379328e7119\") " Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.507931 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b9d91b74-236c-4989-93ca-2379328e7119" (UID: "b9d91b74-236c-4989-93ca-2379328e7119"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.508113 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-logs" (OuterVolumeSpecName: "logs") pod "b9d91b74-236c-4989-93ca-2379328e7119" (UID: "b9d91b74-236c-4989-93ca-2379328e7119"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.508815 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.508830 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b9d91b74-236c-4989-93ca-2379328e7119-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.517783 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "b9d91b74-236c-4989-93ca-2379328e7119" (UID: "b9d91b74-236c-4989-93ca-2379328e7119"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:47:01 crc kubenswrapper[4805]: W1128 15:47:01.539485 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1232c9a3_93bb_4956_8e76_89829ee8b299.slice/crio-0835ab57c7d7770bff0b6de8b5faf934f36c9d159b7381278b65e62b7937989e WatchSource:0}: Error finding container 0835ab57c7d7770bff0b6de8b5faf934f36c9d159b7381278b65e62b7937989e: Status 404 returned error can't find the container with id 0835ab57c7d7770bff0b6de8b5faf934f36c9d159b7381278b65e62b7937989e Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.551492 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-scripts" (OuterVolumeSpecName: "scripts") pod "b9d91b74-236c-4989-93ca-2379328e7119" (UID: "b9d91b74-236c-4989-93ca-2379328e7119"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.561588 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9d91b74-236c-4989-93ca-2379328e7119-kube-api-access-75lzn" (OuterVolumeSpecName: "kube-api-access-75lzn") pod "b9d91b74-236c-4989-93ca-2379328e7119" (UID: "b9d91b74-236c-4989-93ca-2379328e7119"). InnerVolumeSpecName "kube-api-access-75lzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.572437 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9d91b74-236c-4989-93ca-2379328e7119" (UID: "b9d91b74-236c-4989-93ca-2379328e7119"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.610078 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.610170 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.610232 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.610286 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75lzn\" (UniqueName: \"kubernetes.io/projected/b9d91b74-236c-4989-93ca-2379328e7119-kube-api-access-75lzn\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.672436 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wn5s2"] Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.679514 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rr5bn" event={"ID":"3672f82d-1221-47a2-8be9-cbc3394def33","Type":"ContainerStarted","Data":"f8566ba278f7725c827f682fa40e5bc7af932461753c88f412a5530ab9a78ea3"} Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.679570 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rr5bn" event={"ID":"3672f82d-1221-47a2-8be9-cbc3394def33","Type":"ContainerStarted","Data":"3557a6e6ecf9d1e2aa83111da51e315a11e4b82b32f08b4574473a7ab9cae52b"} Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.692419 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54ff5706-18a9-4e71-8289-e3d9e5d9af23","Type":"ContainerStarted","Data":"4f7f7afc6966bf70692a1ec631ffcd04f68c6d162a03e988818a686b37ca27b5"} Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.709999 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8782-account-create-update-p6l42" event={"ID":"2efa09bc-8af5-473d-a893-0cdd3936a6ce","Type":"ContainerStarted","Data":"8ee56f193fc56ad8efaa7bd73c7c776298ba7221ddb1d8dd6f917941d88c46cc"} Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.729956 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.749742 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vkkgt" event={"ID":"1232c9a3-93bb-4956-8e76-89829ee8b299","Type":"ContainerStarted","Data":"0835ab57c7d7770bff0b6de8b5faf934f36c9d159b7381278b65e62b7937989e"} Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.749880 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-config-data" (OuterVolumeSpecName: "config-data") pod "b9d91b74-236c-4989-93ca-2379328e7119" (UID: "b9d91b74-236c-4989-93ca-2379328e7119"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.756053 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b9d91b74-236c-4989-93ca-2379328e7119" (UID: "b9d91b74-236c-4989-93ca-2379328e7119"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:01 crc kubenswrapper[4805]: W1128 15:47:01.757443 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35e934e2_71c2_40ee_bcc7_bbd04ecbac9c.slice/crio-add9020a99f1d105123f51ee613258150ae2fdebba1463322c97989c9e98fca1 WatchSource:0}: Error finding container add9020a99f1d105123f51ee613258150ae2fdebba1463322c97989c9e98fca1: Status 404 returned error can't find the container with id add9020a99f1d105123f51ee613258150ae2fdebba1463322c97989c9e98fca1 Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.762479 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.762505 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b9d91b74-236c-4989-93ca-2379328e7119","Type":"ContainerDied","Data":"e97cb1db4f9760c1fe5620f60e7cc8cc379267d2ac3cea0acf084549a08d85be"} Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.763718 4805 scope.go:117] "RemoveContainer" containerID="1af8fb2da9461341a610764bda29e02ea05cb8bafd6f608ea0e1266702bb14e5" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.776247 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c2b1-account-create-update-qgsb8"] Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.821222 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-rr5bn" podStartSLOduration=2.821206673 podStartE2EDuration="2.821206673s" podCreationTimestamp="2025-11-28 15:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:01.702755518 +0000 UTC m=+1248.752546829" watchObservedRunningTime="2025-11-28 15:47:01.821206673 +0000 UTC m=+1248.870997984" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.831252 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.831671 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d91b74-236c-4989-93ca-2379328e7119-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.832075 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.855448 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1502-account-create-update-2sqhk"] Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.875572 4805 scope.go:117] "RemoveContainer" containerID="c655e1242a0c7987b02a1c84687637804a2b82ea438583fed89c15a2e0f3dc56" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.884808 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.931333 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.960541 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:47:01 crc kubenswrapper[4805]: E1128 15:47:01.961136 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-httpd" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.961147 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-httpd" Nov 28 15:47:01 crc kubenswrapper[4805]: E1128 15:47:01.961168 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-log" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.961174 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-log" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.961405 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-log" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.961421 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9d91b74-236c-4989-93ca-2379328e7119" containerName="glance-httpd" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.962311 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.969621 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.969812 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 15:47:01 crc kubenswrapper[4805]: I1128 15:47:01.971945 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.035402 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-config-data\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.035444 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-scripts\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.035537 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.035565 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.035586 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.035600 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-logs\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.035834 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.035949 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx5gn\" (UniqueName: \"kubernetes.io/projected/19ef9cc1-28ab-4014-897c-9679f3d36443-kube-api-access-vx5gn\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.046758 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:02 crc kubenswrapper[4805]: W1128 15:47:02.066243 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ade9363_30e9_45e1_a57b_00704d8277ca.slice/crio-9d0a8e052d5133c976a37a0b6e1be8701418770ed486226efa79454b58d9e8c4 WatchSource:0}: Error finding container 9d0a8e052d5133c976a37a0b6e1be8701418770ed486226efa79454b58d9e8c4: Status 404 returned error can't find the container with id 9d0a8e052d5133c976a37a0b6e1be8701418770ed486226efa79454b58d9e8c4 Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.137872 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.137937 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.137965 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.137990 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-logs\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.138019 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.139463 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx5gn\" (UniqueName: \"kubernetes.io/projected/19ef9cc1-28ab-4014-897c-9679f3d36443-kube-api-access-vx5gn\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.139646 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-config-data\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.139675 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-scripts\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.140705 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.141486 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.141999 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-logs\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.146469 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.146820 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.147413 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-config-data\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.147440 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-scripts\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.163046 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx5gn\" (UniqueName: \"kubernetes.io/projected/19ef9cc1-28ab-4014-897c-9679f3d36443-kube-api-access-vx5gn\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.183008 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.285863 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.303255 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.484897 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.797426 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerStarted","Data":"9d0a8e052d5133c976a37a0b6e1be8701418770ed486226efa79454b58d9e8c4"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.838542 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54ff5706-18a9-4e71-8289-e3d9e5d9af23","Type":"ContainerStarted","Data":"8007f0c48cb665204284cbdc1681cef5de09b06e0e0fbddc280c197c6fe89921"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.876570 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1502-account-create-update-2sqhk" event={"ID":"79634ed6-4d3f-4488-8123-bce2cb65249d","Type":"ContainerStarted","Data":"52836939c88b31b81ae1b8b1369c201d032ad656c44f7c7999db8fd5105ccd6a"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.876626 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1502-account-create-update-2sqhk" event={"ID":"79634ed6-4d3f-4488-8123-bce2cb65249d","Type":"ContainerStarted","Data":"252b46fcc5caeaddb6a38beaaf5d8ed69e8934402878d219db457cdb05baadff"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.892009 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.891986511 podStartE2EDuration="4.891986511s" podCreationTimestamp="2025-11-28 15:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:02.877853567 +0000 UTC m=+1249.927644878" watchObservedRunningTime="2025-11-28 15:47:02.891986511 +0000 UTC m=+1249.941777832" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.895827 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8782-account-create-update-p6l42" event={"ID":"2efa09bc-8af5-473d-a893-0cdd3936a6ce","Type":"ContainerStarted","Data":"ab13ce4cb62ee031c9d0c066d028e77475c256f8b5848622a9fddf54b469b612"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.906596 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vkkgt" event={"ID":"1232c9a3-93bb-4956-8e76-89829ee8b299","Type":"ContainerStarted","Data":"fe1f555311c706978cf71b633f58b4d74d761a478727bbea477a1a6d7694312a"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.912837 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wn5s2" event={"ID":"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c","Type":"ContainerStarted","Data":"23ec9bca9f3899799fc7a6bd1a69fcbc436496ef980bbce600934b76418db1a3"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.912901 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wn5s2" event={"ID":"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c","Type":"ContainerStarted","Data":"add9020a99f1d105123f51ee613258150ae2fdebba1463322c97989c9e98fca1"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.927041 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c2b1-account-create-update-qgsb8" event={"ID":"d6cb1991-b179-4169-be04-b0e707cc99a3","Type":"ContainerStarted","Data":"eb8fb64ef7fabdfff26fc63df7f5afd8cdf3b2709521bf51427cfd02832aac89"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.927090 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c2b1-account-create-update-qgsb8" event={"ID":"d6cb1991-b179-4169-be04-b0e707cc99a3","Type":"ContainerStarted","Data":"b3b2c20c4acd6dcb37e08afb9186c76b706c200770cfa6356b06e90e6d91cfd7"} Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.942465 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.943271 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-1502-account-create-update-2sqhk" podStartSLOduration=3.943256507 podStartE2EDuration="3.943256507s" podCreationTimestamp="2025-11-28 15:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:02.901768267 +0000 UTC m=+1249.951559588" watchObservedRunningTime="2025-11-28 15:47:02.943256507 +0000 UTC m=+1249.993047808" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.973633 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-8782-account-create-update-p6l42" podStartSLOduration=3.973614983 podStartE2EDuration="3.973614983s" podCreationTimestamp="2025-11-28 15:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:02.923101068 +0000 UTC m=+1249.972892389" watchObservedRunningTime="2025-11-28 15:47:02.973614983 +0000 UTC m=+1250.023406294" Nov 28 15:47:02 crc kubenswrapper[4805]: I1128 15:47:02.995872 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-wn5s2" podStartSLOduration=3.9957758070000002 podStartE2EDuration="3.995775807s" podCreationTimestamp="2025-11-28 15:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:02.941061157 +0000 UTC m=+1249.990852468" watchObservedRunningTime="2025-11-28 15:47:02.995775807 +0000 UTC m=+1250.045567128" Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.029203 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-vkkgt" podStartSLOduration=4.029179166 podStartE2EDuration="4.029179166s" podCreationTimestamp="2025-11-28 15:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:02.954752229 +0000 UTC m=+1250.004543550" watchObservedRunningTime="2025-11-28 15:47:03.029179166 +0000 UTC m=+1250.078970477" Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.120049 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-c2b1-account-create-update-qgsb8" podStartSLOduration=4.120027839 podStartE2EDuration="4.120027839s" podCreationTimestamp="2025-11-28 15:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:02.975733951 +0000 UTC m=+1250.025525262" watchObservedRunningTime="2025-11-28 15:47:03.120027839 +0000 UTC m=+1250.169819150" Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.242941 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9d91b74-236c-4989-93ca-2379328e7119" path="/var/lib/kubelet/pods/b9d91b74-236c-4989-93ca-2379328e7119/volumes" Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.940612 4805 generic.go:334] "Generic (PLEG): container finished" podID="1232c9a3-93bb-4956-8e76-89829ee8b299" containerID="fe1f555311c706978cf71b633f58b4d74d761a478727bbea477a1a6d7694312a" exitCode=0 Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.940660 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vkkgt" event={"ID":"1232c9a3-93bb-4956-8e76-89829ee8b299","Type":"ContainerDied","Data":"fe1f555311c706978cf71b633f58b4d74d761a478727bbea477a1a6d7694312a"} Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.945302 4805 generic.go:334] "Generic (PLEG): container finished" podID="d6cb1991-b179-4169-be04-b0e707cc99a3" containerID="eb8fb64ef7fabdfff26fc63df7f5afd8cdf3b2709521bf51427cfd02832aac89" exitCode=0 Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.947065 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c2b1-account-create-update-qgsb8" event={"ID":"d6cb1991-b179-4169-be04-b0e707cc99a3","Type":"ContainerDied","Data":"eb8fb64ef7fabdfff26fc63df7f5afd8cdf3b2709521bf51427cfd02832aac89"} Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.957393 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"19ef9cc1-28ab-4014-897c-9679f3d36443","Type":"ContainerStarted","Data":"4d900ec2e1183b4ce27e8479bb7bc8fa6d050fc7702be965871220be9e57e825"} Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.957457 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"19ef9cc1-28ab-4014-897c-9679f3d36443","Type":"ContainerStarted","Data":"93d4d2656c06ae1edfa165df51ed76d714acbc02dde56fdfceb090c51aeebe72"} Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.959305 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerStarted","Data":"d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579"} Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.965359 4805 generic.go:334] "Generic (PLEG): container finished" podID="3672f82d-1221-47a2-8be9-cbc3394def33" containerID="f8566ba278f7725c827f682fa40e5bc7af932461753c88f412a5530ab9a78ea3" exitCode=0 Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.965469 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rr5bn" event={"ID":"3672f82d-1221-47a2-8be9-cbc3394def33","Type":"ContainerDied","Data":"f8566ba278f7725c827f682fa40e5bc7af932461753c88f412a5530ab9a78ea3"} Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.971216 4805 generic.go:334] "Generic (PLEG): container finished" podID="79634ed6-4d3f-4488-8123-bce2cb65249d" containerID="52836939c88b31b81ae1b8b1369c201d032ad656c44f7c7999db8fd5105ccd6a" exitCode=0 Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.971268 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1502-account-create-update-2sqhk" event={"ID":"79634ed6-4d3f-4488-8123-bce2cb65249d","Type":"ContainerDied","Data":"52836939c88b31b81ae1b8b1369c201d032ad656c44f7c7999db8fd5105ccd6a"} Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.980253 4805 generic.go:334] "Generic (PLEG): container finished" podID="2efa09bc-8af5-473d-a893-0cdd3936a6ce" containerID="ab13ce4cb62ee031c9d0c066d028e77475c256f8b5848622a9fddf54b469b612" exitCode=0 Nov 28 15:47:03 crc kubenswrapper[4805]: I1128 15:47:03.980347 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8782-account-create-update-p6l42" event={"ID":"2efa09bc-8af5-473d-a893-0cdd3936a6ce","Type":"ContainerDied","Data":"ab13ce4cb62ee031c9d0c066d028e77475c256f8b5848622a9fddf54b469b612"} Nov 28 15:47:04 crc kubenswrapper[4805]: I1128 15:47:04.004756 4805 generic.go:334] "Generic (PLEG): container finished" podID="35e934e2-71c2-40ee-bcc7-bbd04ecbac9c" containerID="23ec9bca9f3899799fc7a6bd1a69fcbc436496ef980bbce600934b76418db1a3" exitCode=0 Nov 28 15:47:04 crc kubenswrapper[4805]: I1128 15:47:04.005969 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wn5s2" event={"ID":"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c","Type":"ContainerDied","Data":"23ec9bca9f3899799fc7a6bd1a69fcbc436496ef980bbce600934b76418db1a3"} Nov 28 15:47:04 crc kubenswrapper[4805]: I1128 15:47:04.152982 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.033904 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"19ef9cc1-28ab-4014-897c-9679f3d36443","Type":"ContainerStarted","Data":"868c845c22044f953c99bf70b59a10a7cf406a71b3f9ce6c6262250153d75b30"} Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.036447 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerStarted","Data":"ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974"} Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.072036 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.072010985 podStartE2EDuration="4.072010985s" podCreationTimestamp="2025-11-28 15:47:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:05.065071467 +0000 UTC m=+1252.114862768" watchObservedRunningTime="2025-11-28 15:47:05.072010985 +0000 UTC m=+1252.121802296" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.622121 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.663141 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.739755 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efa09bc-8af5-473d-a893-0cdd3936a6ce-operator-scripts\") pod \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\" (UID: \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\") " Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.739816 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcpp6\" (UniqueName: \"kubernetes.io/projected/2efa09bc-8af5-473d-a893-0cdd3936a6ce-kube-api-access-lcpp6\") pod \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\" (UID: \"2efa09bc-8af5-473d-a893-0cdd3936a6ce\") " Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.741953 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2efa09bc-8af5-473d-a893-0cdd3936a6ce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2efa09bc-8af5-473d-a893-0cdd3936a6ce" (UID: "2efa09bc-8af5-473d-a893-0cdd3936a6ce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.759880 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-79b56c885b-5nx79"] Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.760162 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-79b56c885b-5nx79" podUID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerName="neutron-api" containerID="cri-o://904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147" gracePeriod=30 Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.761226 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-79b56c885b-5nx79" podUID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerName="neutron-httpd" containerID="cri-o://2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee" gracePeriod=30 Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.775561 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2efa09bc-8af5-473d-a893-0cdd3936a6ce-kube-api-access-lcpp6" (OuterVolumeSpecName: "kube-api-access-lcpp6") pod "2efa09bc-8af5-473d-a893-0cdd3936a6ce" (UID: "2efa09bc-8af5-473d-a893-0cdd3936a6ce"). InnerVolumeSpecName "kube-api-access-lcpp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.845344 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efa09bc-8af5-473d-a893-0cdd3936a6ce-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.845398 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcpp6\" (UniqueName: \"kubernetes.io/projected/2efa09bc-8af5-473d-a893-0cdd3936a6ce-kube-api-access-lcpp6\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.846266 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.879744 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.948558 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2rl2\" (UniqueName: \"kubernetes.io/projected/79634ed6-4d3f-4488-8123-bce2cb65249d-kube-api-access-w2rl2\") pod \"79634ed6-4d3f-4488-8123-bce2cb65249d\" (UID: \"79634ed6-4d3f-4488-8123-bce2cb65249d\") " Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.948631 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79634ed6-4d3f-4488-8123-bce2cb65249d-operator-scripts\") pod \"79634ed6-4d3f-4488-8123-bce2cb65249d\" (UID: \"79634ed6-4d3f-4488-8123-bce2cb65249d\") " Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.948793 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w424v\" (UniqueName: \"kubernetes.io/projected/1232c9a3-93bb-4956-8e76-89829ee8b299-kube-api-access-w424v\") pod \"1232c9a3-93bb-4956-8e76-89829ee8b299\" (UID: \"1232c9a3-93bb-4956-8e76-89829ee8b299\") " Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.948811 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1232c9a3-93bb-4956-8e76-89829ee8b299-operator-scripts\") pod \"1232c9a3-93bb-4956-8e76-89829ee8b299\" (UID: \"1232c9a3-93bb-4956-8e76-89829ee8b299\") " Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.949632 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1232c9a3-93bb-4956-8e76-89829ee8b299-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1232c9a3-93bb-4956-8e76-89829ee8b299" (UID: "1232c9a3-93bb-4956-8e76-89829ee8b299"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.949636 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79634ed6-4d3f-4488-8123-bce2cb65249d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "79634ed6-4d3f-4488-8123-bce2cb65249d" (UID: "79634ed6-4d3f-4488-8123-bce2cb65249d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.963587 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79634ed6-4d3f-4488-8123-bce2cb65249d-kube-api-access-w2rl2" (OuterVolumeSpecName: "kube-api-access-w2rl2") pod "79634ed6-4d3f-4488-8123-bce2cb65249d" (UID: "79634ed6-4d3f-4488-8123-bce2cb65249d"). InnerVolumeSpecName "kube-api-access-w2rl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.963685 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1232c9a3-93bb-4956-8e76-89829ee8b299-kube-api-access-w424v" (OuterVolumeSpecName: "kube-api-access-w424v") pod "1232c9a3-93bb-4956-8e76-89829ee8b299" (UID: "1232c9a3-93bb-4956-8e76-89829ee8b299"). InnerVolumeSpecName "kube-api-access-w424v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.964001 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:47:05 crc kubenswrapper[4805]: I1128 15:47:05.983561 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.053189 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3672f82d-1221-47a2-8be9-cbc3394def33-operator-scripts\") pod \"3672f82d-1221-47a2-8be9-cbc3394def33\" (UID: \"3672f82d-1221-47a2-8be9-cbc3394def33\") " Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.053596 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzbb6\" (UniqueName: \"kubernetes.io/projected/3672f82d-1221-47a2-8be9-cbc3394def33-kube-api-access-qzbb6\") pod \"3672f82d-1221-47a2-8be9-cbc3394def33\" (UID: \"3672f82d-1221-47a2-8be9-cbc3394def33\") " Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.053631 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nq6j\" (UniqueName: \"kubernetes.io/projected/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-kube-api-access-9nq6j\") pod \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\" (UID: \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\") " Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.053698 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-operator-scripts\") pod \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\" (UID: \"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c\") " Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.054025 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3672f82d-1221-47a2-8be9-cbc3394def33-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3672f82d-1221-47a2-8be9-cbc3394def33" (UID: "3672f82d-1221-47a2-8be9-cbc3394def33"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.054078 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w424v\" (UniqueName: \"kubernetes.io/projected/1232c9a3-93bb-4956-8e76-89829ee8b299-kube-api-access-w424v\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.054090 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1232c9a3-93bb-4956-8e76-89829ee8b299-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.054098 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2rl2\" (UniqueName: \"kubernetes.io/projected/79634ed6-4d3f-4488-8123-bce2cb65249d-kube-api-access-w2rl2\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.054107 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79634ed6-4d3f-4488-8123-bce2cb65249d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.054512 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "35e934e2-71c2-40ee-bcc7-bbd04ecbac9c" (UID: "35e934e2-71c2-40ee-bcc7-bbd04ecbac9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.068599 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-kube-api-access-9nq6j" (OuterVolumeSpecName: "kube-api-access-9nq6j") pod "35e934e2-71c2-40ee-bcc7-bbd04ecbac9c" (UID: "35e934e2-71c2-40ee-bcc7-bbd04ecbac9c"). InnerVolumeSpecName "kube-api-access-9nq6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.083737 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3672f82d-1221-47a2-8be9-cbc3394def33-kube-api-access-qzbb6" (OuterVolumeSpecName: "kube-api-access-qzbb6") pod "3672f82d-1221-47a2-8be9-cbc3394def33" (UID: "3672f82d-1221-47a2-8be9-cbc3394def33"). InnerVolumeSpecName "kube-api-access-qzbb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.093407 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerStarted","Data":"ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76"} Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.112423 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rr5bn" event={"ID":"3672f82d-1221-47a2-8be9-cbc3394def33","Type":"ContainerDied","Data":"3557a6e6ecf9d1e2aa83111da51e315a11e4b82b32f08b4574473a7ab9cae52b"} Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.112460 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3557a6e6ecf9d1e2aa83111da51e315a11e4b82b32f08b4574473a7ab9cae52b" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.112508 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rr5bn" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.131829 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1502-account-create-update-2sqhk" event={"ID":"79634ed6-4d3f-4488-8123-bce2cb65249d","Type":"ContainerDied","Data":"252b46fcc5caeaddb6a38beaaf5d8ed69e8934402878d219db457cdb05baadff"} Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.131878 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="252b46fcc5caeaddb6a38beaaf5d8ed69e8934402878d219db457cdb05baadff" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.131969 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1502-account-create-update-2sqhk" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.137010 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.137687 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8782-account-create-update-p6l42" event={"ID":"2efa09bc-8af5-473d-a893-0cdd3936a6ce","Type":"ContainerDied","Data":"8ee56f193fc56ad8efaa7bd73c7c776298ba7221ddb1d8dd6f917941d88c46cc"} Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.137709 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ee56f193fc56ad8efaa7bd73c7c776298ba7221ddb1d8dd6f917941d88c46cc" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.137752 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8782-account-create-update-p6l42" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.156323 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzbb6\" (UniqueName: \"kubernetes.io/projected/3672f82d-1221-47a2-8be9-cbc3394def33-kube-api-access-qzbb6\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.156355 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nq6j\" (UniqueName: \"kubernetes.io/projected/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-kube-api-access-9nq6j\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.156389 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.156398 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3672f82d-1221-47a2-8be9-cbc3394def33-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.164939 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vkkgt" event={"ID":"1232c9a3-93bb-4956-8e76-89829ee8b299","Type":"ContainerDied","Data":"0835ab57c7d7770bff0b6de8b5faf934f36c9d159b7381278b65e62b7937989e"} Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.164977 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0835ab57c7d7770bff0b6de8b5faf934f36c9d159b7381278b65e62b7937989e" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.165022 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vkkgt" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.170116 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wn5s2" event={"ID":"35e934e2-71c2-40ee-bcc7-bbd04ecbac9c","Type":"ContainerDied","Data":"add9020a99f1d105123f51ee613258150ae2fdebba1463322c97989c9e98fca1"} Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.170163 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="add9020a99f1d105123f51ee613258150ae2fdebba1463322c97989c9e98fca1" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.170229 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wn5s2" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.180097 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c2b1-account-create-update-qgsb8" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.180230 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c2b1-account-create-update-qgsb8" event={"ID":"d6cb1991-b179-4169-be04-b0e707cc99a3","Type":"ContainerDied","Data":"b3b2c20c4acd6dcb37e08afb9186c76b706c200770cfa6356b06e90e6d91cfd7"} Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.180249 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3b2c20c4acd6dcb37e08afb9186c76b706c200770cfa6356b06e90e6d91cfd7" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.263927 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6njc8\" (UniqueName: \"kubernetes.io/projected/d6cb1991-b179-4169-be04-b0e707cc99a3-kube-api-access-6njc8\") pod \"d6cb1991-b179-4169-be04-b0e707cc99a3\" (UID: \"d6cb1991-b179-4169-be04-b0e707cc99a3\") " Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.264049 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6cb1991-b179-4169-be04-b0e707cc99a3-operator-scripts\") pod \"d6cb1991-b179-4169-be04-b0e707cc99a3\" (UID: \"d6cb1991-b179-4169-be04-b0e707cc99a3\") " Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.266649 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6cb1991-b179-4169-be04-b0e707cc99a3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d6cb1991-b179-4169-be04-b0e707cc99a3" (UID: "d6cb1991-b179-4169-be04-b0e707cc99a3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.271254 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6cb1991-b179-4169-be04-b0e707cc99a3-kube-api-access-6njc8" (OuterVolumeSpecName: "kube-api-access-6njc8") pod "d6cb1991-b179-4169-be04-b0e707cc99a3" (UID: "d6cb1991-b179-4169-be04-b0e707cc99a3"). InnerVolumeSpecName "kube-api-access-6njc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.366656 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6njc8\" (UniqueName: \"kubernetes.io/projected/d6cb1991-b179-4169-be04-b0e707cc99a3-kube-api-access-6njc8\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.366694 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6cb1991-b179-4169-be04-b0e707cc99a3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.475016 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.475459 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerName="glance-httpd" containerID="cri-o://b050f44a1b2eee2ec7d7f2eaa9e0af5107a6d79deecade5b296d226400c82649" gracePeriod=30 Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.475461 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerName="glance-log" containerID="cri-o://3132a6f16b1d3aee6579746b5437f124996d0c2d9789b01ebda38770a8012732" gracePeriod=30 Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.955713 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:47:06 crc kubenswrapper[4805]: I1128 15:47:06.963215 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:47:07 crc kubenswrapper[4805]: I1128 15:47:07.194598 4805 generic.go:334] "Generic (PLEG): container finished" podID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerID="2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee" exitCode=0 Nov 28 15:47:07 crc kubenswrapper[4805]: I1128 15:47:07.194980 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b56c885b-5nx79" event={"ID":"764e5362-53f7-46e3-a31b-5a0c679f3484","Type":"ContainerDied","Data":"2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee"} Nov 28 15:47:07 crc kubenswrapper[4805]: I1128 15:47:07.196811 4805 generic.go:334] "Generic (PLEG): container finished" podID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerID="3132a6f16b1d3aee6579746b5437f124996d0c2d9789b01ebda38770a8012732" exitCode=143 Nov 28 15:47:07 crc kubenswrapper[4805]: I1128 15:47:07.196862 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17e2c4d9-4960-4034-be79-5dffbab66c5b","Type":"ContainerDied","Data":"3132a6f16b1d3aee6579746b5437f124996d0c2d9789b01ebda38770a8012732"} Nov 28 15:47:09 crc kubenswrapper[4805]: I1128 15:47:09.234744 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerStarted","Data":"98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364"} Nov 28 15:47:09 crc kubenswrapper[4805]: I1128 15:47:09.237021 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 15:47:09 crc kubenswrapper[4805]: I1128 15:47:09.235603 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="proxy-httpd" containerID="cri-o://98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364" gracePeriod=30 Nov 28 15:47:09 crc kubenswrapper[4805]: I1128 15:47:09.235051 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="ceilometer-central-agent" containerID="cri-o://d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579" gracePeriod=30 Nov 28 15:47:09 crc kubenswrapper[4805]: I1128 15:47:09.235634 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="ceilometer-notification-agent" containerID="cri-o://ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974" gracePeriod=30 Nov 28 15:47:09 crc kubenswrapper[4805]: I1128 15:47:09.235623 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="sg-core" containerID="cri-o://ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76" gracePeriod=30 Nov 28 15:47:09 crc kubenswrapper[4805]: I1128 15:47:09.268717 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.305595179 podStartE2EDuration="9.268700376s" podCreationTimestamp="2025-11-28 15:47:00 +0000 UTC" firstStartedPulling="2025-11-28 15:47:02.077497389 +0000 UTC m=+1249.127288700" lastFinishedPulling="2025-11-28 15:47:08.040602586 +0000 UTC m=+1255.090393897" observedRunningTime="2025-11-28 15:47:09.260257546 +0000 UTC m=+1256.310048857" watchObservedRunningTime="2025-11-28 15:47:09.268700376 +0000 UTC m=+1256.318491677" Nov 28 15:47:09 crc kubenswrapper[4805]: I1128 15:47:09.374653 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.079481 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cfqg8"] Nov 28 15:47:10 crc kubenswrapper[4805]: E1128 15:47:10.080098 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e934e2-71c2-40ee-bcc7-bbd04ecbac9c" containerName="mariadb-database-create" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080114 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e934e2-71c2-40ee-bcc7-bbd04ecbac9c" containerName="mariadb-database-create" Nov 28 15:47:10 crc kubenswrapper[4805]: E1128 15:47:10.080137 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1232c9a3-93bb-4956-8e76-89829ee8b299" containerName="mariadb-database-create" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080144 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1232c9a3-93bb-4956-8e76-89829ee8b299" containerName="mariadb-database-create" Nov 28 15:47:10 crc kubenswrapper[4805]: E1128 15:47:10.080160 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6cb1991-b179-4169-be04-b0e707cc99a3" containerName="mariadb-account-create-update" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080166 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6cb1991-b179-4169-be04-b0e707cc99a3" containerName="mariadb-account-create-update" Nov 28 15:47:10 crc kubenswrapper[4805]: E1128 15:47:10.080175 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3672f82d-1221-47a2-8be9-cbc3394def33" containerName="mariadb-database-create" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080181 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="3672f82d-1221-47a2-8be9-cbc3394def33" containerName="mariadb-database-create" Nov 28 15:47:10 crc kubenswrapper[4805]: E1128 15:47:10.080198 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79634ed6-4d3f-4488-8123-bce2cb65249d" containerName="mariadb-account-create-update" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080204 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="79634ed6-4d3f-4488-8123-bce2cb65249d" containerName="mariadb-account-create-update" Nov 28 15:47:10 crc kubenswrapper[4805]: E1128 15:47:10.080217 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2efa09bc-8af5-473d-a893-0cdd3936a6ce" containerName="mariadb-account-create-update" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080223 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2efa09bc-8af5-473d-a893-0cdd3936a6ce" containerName="mariadb-account-create-update" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080399 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e934e2-71c2-40ee-bcc7-bbd04ecbac9c" containerName="mariadb-database-create" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080414 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="3672f82d-1221-47a2-8be9-cbc3394def33" containerName="mariadb-database-create" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080421 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6cb1991-b179-4169-be04-b0e707cc99a3" containerName="mariadb-account-create-update" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080437 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="79634ed6-4d3f-4488-8123-bce2cb65249d" containerName="mariadb-account-create-update" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080450 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1232c9a3-93bb-4956-8e76-89829ee8b299" containerName="mariadb-database-create" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.080462 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2efa09bc-8af5-473d-a893-0cdd3936a6ce" containerName="mariadb-account-create-update" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.081042 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.084602 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-7gz9m" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.084695 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.084621 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.092487 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cfqg8"] Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.144575 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9fh6\" (UniqueName: \"kubernetes.io/projected/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-kube-api-access-r9fh6\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.144619 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-scripts\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.144654 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-config-data\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.144711 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.168624 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.245418 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-combined-ca-bundle\") pod \"764e5362-53f7-46e3-a31b-5a0c679f3484\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.245570 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-httpd-config\") pod \"764e5362-53f7-46e3-a31b-5a0c679f3484\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.245624 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-config\") pod \"764e5362-53f7-46e3-a31b-5a0c679f3484\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.245735 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-ovndb-tls-certs\") pod \"764e5362-53f7-46e3-a31b-5a0c679f3484\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.245808 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwm28\" (UniqueName: \"kubernetes.io/projected/764e5362-53f7-46e3-a31b-5a0c679f3484-kube-api-access-nwm28\") pod \"764e5362-53f7-46e3-a31b-5a0c679f3484\" (UID: \"764e5362-53f7-46e3-a31b-5a0c679f3484\") " Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.246067 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9fh6\" (UniqueName: \"kubernetes.io/projected/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-kube-api-access-r9fh6\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.246089 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-scripts\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.246120 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-config-data\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.246185 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.250462 4805 generic.go:334] "Generic (PLEG): container finished" podID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerID="98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364" exitCode=0 Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.250496 4805 generic.go:334] "Generic (PLEG): container finished" podID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerID="ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76" exitCode=2 Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.250505 4805 generic.go:334] "Generic (PLEG): container finished" podID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerID="ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974" exitCode=0 Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.250547 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerDied","Data":"98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364"} Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.250573 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerDied","Data":"ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76"} Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.250583 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerDied","Data":"ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974"} Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.252808 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/764e5362-53f7-46e3-a31b-5a0c679f3484-kube-api-access-nwm28" (OuterVolumeSpecName: "kube-api-access-nwm28") pod "764e5362-53f7-46e3-a31b-5a0c679f3484" (UID: "764e5362-53f7-46e3-a31b-5a0c679f3484"). InnerVolumeSpecName "kube-api-access-nwm28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.252939 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.253243 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "764e5362-53f7-46e3-a31b-5a0c679f3484" (UID: "764e5362-53f7-46e3-a31b-5a0c679f3484"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.253685 4805 generic.go:334] "Generic (PLEG): container finished" podID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerID="904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147" exitCode=0 Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.253812 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79b56c885b-5nx79" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.254490 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b56c885b-5nx79" event={"ID":"764e5362-53f7-46e3-a31b-5a0c679f3484","Type":"ContainerDied","Data":"904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147"} Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.254522 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b56c885b-5nx79" event={"ID":"764e5362-53f7-46e3-a31b-5a0c679f3484","Type":"ContainerDied","Data":"7e998d046840d741e1be77452ff2eba6aeae6eb242cecd212ab0c168441e4b42"} Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.254540 4805 scope.go:117] "RemoveContainer" containerID="2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.255001 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-config-data\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.255909 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-scripts\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.259270 4805 generic.go:334] "Generic (PLEG): container finished" podID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerID="b050f44a1b2eee2ec7d7f2eaa9e0af5107a6d79deecade5b296d226400c82649" exitCode=0 Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.259299 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17e2c4d9-4960-4034-be79-5dffbab66c5b","Type":"ContainerDied","Data":"b050f44a1b2eee2ec7d7f2eaa9e0af5107a6d79deecade5b296d226400c82649"} Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.266787 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9fh6\" (UniqueName: \"kubernetes.io/projected/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-kube-api-access-r9fh6\") pod \"nova-cell0-conductor-db-sync-cfqg8\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.285495 4805 scope.go:117] "RemoveContainer" containerID="904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.300564 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "764e5362-53f7-46e3-a31b-5a0c679f3484" (UID: "764e5362-53f7-46e3-a31b-5a0c679f3484"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.303458 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-config" (OuterVolumeSpecName: "config") pod "764e5362-53f7-46e3-a31b-5a0c679f3484" (UID: "764e5362-53f7-46e3-a31b-5a0c679f3484"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.310770 4805 scope.go:117] "RemoveContainer" containerID="2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee" Nov 28 15:47:10 crc kubenswrapper[4805]: E1128 15:47:10.311345 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee\": container with ID starting with 2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee not found: ID does not exist" containerID="2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.311421 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee"} err="failed to get container status \"2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee\": rpc error: code = NotFound desc = could not find container \"2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee\": container with ID starting with 2f89d81665eca812c33585c437a5a1c25d6dbe00731ad4f5bdee97fe5ff49eee not found: ID does not exist" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.311453 4805 scope.go:117] "RemoveContainer" containerID="904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147" Nov 28 15:47:10 crc kubenswrapper[4805]: E1128 15:47:10.312118 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147\": container with ID starting with 904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147 not found: ID does not exist" containerID="904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.312167 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147"} err="failed to get container status \"904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147\": rpc error: code = NotFound desc = could not find container \"904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147\": container with ID starting with 904dbb026f385554e0428d653c5aa6c3b7a8cfc79a8712b61aa8f27faff7e147 not found: ID does not exist" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.333461 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "764e5362-53f7-46e3-a31b-5a0c679f3484" (UID: "764e5362-53f7-46e3-a31b-5a0c679f3484"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.350274 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwm28\" (UniqueName: \"kubernetes.io/projected/764e5362-53f7-46e3-a31b-5a0c679f3484-kube-api-access-nwm28\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.350308 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.350318 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.350327 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.350337 4805 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/764e5362-53f7-46e3-a31b-5a0c679f3484-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.407655 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.595531 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-79b56c885b-5nx79"] Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.608464 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-79b56c885b-5nx79"] Nov 28 15:47:10 crc kubenswrapper[4805]: I1128 15:47:10.942001 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cfqg8"] Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.029173 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.073207 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.073259 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.174645 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-logs\") pod \"17e2c4d9-4960-4034-be79-5dffbab66c5b\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.174712 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-config-data\") pod \"17e2c4d9-4960-4034-be79-5dffbab66c5b\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.174768 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z2kz\" (UniqueName: \"kubernetes.io/projected/17e2c4d9-4960-4034-be79-5dffbab66c5b-kube-api-access-7z2kz\") pod \"17e2c4d9-4960-4034-be79-5dffbab66c5b\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.174798 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"17e2c4d9-4960-4034-be79-5dffbab66c5b\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.174859 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-httpd-run\") pod \"17e2c4d9-4960-4034-be79-5dffbab66c5b\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.174886 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-scripts\") pod \"17e2c4d9-4960-4034-be79-5dffbab66c5b\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.174907 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-combined-ca-bundle\") pod \"17e2c4d9-4960-4034-be79-5dffbab66c5b\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.175009 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-public-tls-certs\") pod \"17e2c4d9-4960-4034-be79-5dffbab66c5b\" (UID: \"17e2c4d9-4960-4034-be79-5dffbab66c5b\") " Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.175458 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "17e2c4d9-4960-4034-be79-5dffbab66c5b" (UID: "17e2c4d9-4960-4034-be79-5dffbab66c5b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.175776 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-logs" (OuterVolumeSpecName: "logs") pod "17e2c4d9-4960-4034-be79-5dffbab66c5b" (UID: "17e2c4d9-4960-4034-be79-5dffbab66c5b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.180763 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17e2c4d9-4960-4034-be79-5dffbab66c5b-kube-api-access-7z2kz" (OuterVolumeSpecName: "kube-api-access-7z2kz") pod "17e2c4d9-4960-4034-be79-5dffbab66c5b" (UID: "17e2c4d9-4960-4034-be79-5dffbab66c5b"). InnerVolumeSpecName "kube-api-access-7z2kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.181227 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-scripts" (OuterVolumeSpecName: "scripts") pod "17e2c4d9-4960-4034-be79-5dffbab66c5b" (UID: "17e2c4d9-4960-4034-be79-5dffbab66c5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.181512 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "17e2c4d9-4960-4034-be79-5dffbab66c5b" (UID: "17e2c4d9-4960-4034-be79-5dffbab66c5b"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.206044 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17e2c4d9-4960-4034-be79-5dffbab66c5b" (UID: "17e2c4d9-4960-4034-be79-5dffbab66c5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.242826 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="764e5362-53f7-46e3-a31b-5a0c679f3484" path="/var/lib/kubelet/pods/764e5362-53f7-46e3-a31b-5a0c679f3484/volumes" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.254058 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "17e2c4d9-4960-4034-be79-5dffbab66c5b" (UID: "17e2c4d9-4960-4034-be79-5dffbab66c5b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.270137 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-config-data" (OuterVolumeSpecName: "config-data") pod "17e2c4d9-4960-4034-be79-5dffbab66c5b" (UID: "17e2c4d9-4960-4034-be79-5dffbab66c5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.273344 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cfqg8" event={"ID":"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6","Type":"ContainerStarted","Data":"26625d56ecaa27a9ae9e5f970565463dd5d522366994d0b1f08be2e4a28548f5"} Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.275667 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17e2c4d9-4960-4034-be79-5dffbab66c5b","Type":"ContainerDied","Data":"a9d59ed5f828699fd3b23a41a47fc4b3d4c4e17931ca38678b23a5d7a1e55fda"} Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.275717 4805 scope.go:117] "RemoveContainer" containerID="b050f44a1b2eee2ec7d7f2eaa9e0af5107a6d79deecade5b296d226400c82649" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.275847 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.277304 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.277331 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.277342 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.277367 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z2kz\" (UniqueName: \"kubernetes.io/projected/17e2c4d9-4960-4034-be79-5dffbab66c5b-kube-api-access-7z2kz\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.277394 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.277405 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17e2c4d9-4960-4034-be79-5dffbab66c5b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.277415 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.277425 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e2c4d9-4960-4034-be79-5dffbab66c5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.295602 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.312713 4805 scope.go:117] "RemoveContainer" containerID="3132a6f16b1d3aee6579746b5437f124996d0c2d9789b01ebda38770a8012732" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.313213 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.320602 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.354546 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:47:11 crc kubenswrapper[4805]: E1128 15:47:11.355111 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerName="glance-log" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.355184 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerName="glance-log" Nov 28 15:47:11 crc kubenswrapper[4805]: E1128 15:47:11.355254 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerName="glance-httpd" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.355313 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerName="glance-httpd" Nov 28 15:47:11 crc kubenswrapper[4805]: E1128 15:47:11.355393 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerName="neutron-httpd" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.355456 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerName="neutron-httpd" Nov 28 15:47:11 crc kubenswrapper[4805]: E1128 15:47:11.355513 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerName="neutron-api" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.355568 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerName="neutron-api" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.355810 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerName="neutron-httpd" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.355904 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerName="glance-log" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.355973 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="764e5362-53f7-46e3-a31b-5a0c679f3484" containerName="neutron-api" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.356036 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e2c4d9-4960-4034-be79-5dffbab66c5b" containerName="glance-httpd" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.357027 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.364774 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.365156 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.379563 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.388476 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.480929 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-scripts\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.481008 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.481035 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.481222 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.481374 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-logs\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.481430 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-config-data\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.481585 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.481643 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqvzt\" (UniqueName: \"kubernetes.io/projected/5c021413-239d-4ba0-979d-1c4b25d5093c-kube-api-access-dqvzt\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.583536 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.583611 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-logs\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.583642 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-config-data\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.583705 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.583739 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqvzt\" (UniqueName: \"kubernetes.io/projected/5c021413-239d-4ba0-979d-1c4b25d5093c-kube-api-access-dqvzt\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.583786 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-scripts\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.583816 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.583852 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.584289 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-logs\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.584303 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.584811 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.587783 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-scripts\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.591009 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.591122 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.591436 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-config-data\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.600921 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqvzt\" (UniqueName: \"kubernetes.io/projected/5c021413-239d-4ba0-979d-1c4b25d5093c-kube-api-access-dqvzt\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.642674 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " pod="openstack/glance-default-external-api-0" Nov 28 15:47:11 crc kubenswrapper[4805]: I1128 15:47:11.740346 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:47:12 crc kubenswrapper[4805]: I1128 15:47:12.278316 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:47:12 crc kubenswrapper[4805]: I1128 15:47:12.303483 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:12 crc kubenswrapper[4805]: I1128 15:47:12.303538 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:12 crc kubenswrapper[4805]: I1128 15:47:12.340407 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:12 crc kubenswrapper[4805]: I1128 15:47:12.359316 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:13 crc kubenswrapper[4805]: I1128 15:47:13.224856 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17e2c4d9-4960-4034-be79-5dffbab66c5b" path="/var/lib/kubelet/pods/17e2c4d9-4960-4034-be79-5dffbab66c5b/volumes" Nov 28 15:47:13 crc kubenswrapper[4805]: I1128 15:47:13.307003 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c021413-239d-4ba0-979d-1c4b25d5093c","Type":"ContainerStarted","Data":"5290a7c9695a774d3cd36d80cc41c5b6f08e321f90789c87b7e2f50f6f5d2099"} Nov 28 15:47:13 crc kubenswrapper[4805]: I1128 15:47:13.307041 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c021413-239d-4ba0-979d-1c4b25d5093c","Type":"ContainerStarted","Data":"8492df5731462df2cd978bdbc1c0b2d7c30e2f8a9a1fa8380cd034b50491f736"} Nov 28 15:47:13 crc kubenswrapper[4805]: I1128 15:47:13.307058 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:13 crc kubenswrapper[4805]: I1128 15:47:13.307164 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.275173 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.326075 4805 generic.go:334] "Generic (PLEG): container finished" podID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerID="d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579" exitCode=0 Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.326185 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerDied","Data":"d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579"} Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.326215 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ade9363-30e9-45e1-a57b-00704d8277ca","Type":"ContainerDied","Data":"9d0a8e052d5133c976a37a0b6e1be8701418770ed486226efa79454b58d9e8c4"} Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.326241 4805 scope.go:117] "RemoveContainer" containerID="98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.326375 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.329252 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c021413-239d-4ba0-979d-1c4b25d5093c","Type":"ContainerStarted","Data":"31105a43fdab9daa89f1ca09a13406cb4b8904817a89036a2573227173287ae5"} Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.348818 4805 scope.go:117] "RemoveContainer" containerID="ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.366200 4805 scope.go:117] "RemoveContainer" containerID="ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.403811 4805 scope.go:117] "RemoveContainer" containerID="d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.446048 4805 scope.go:117] "RemoveContainer" containerID="98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364" Nov 28 15:47:14 crc kubenswrapper[4805]: E1128 15:47:14.446520 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364\": container with ID starting with 98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364 not found: ID does not exist" containerID="98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.446560 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364"} err="failed to get container status \"98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364\": rpc error: code = NotFound desc = could not find container \"98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364\": container with ID starting with 98421d5c52a8dfe37bd3c409067e6201a489a01d67ad4d0f03e2dce6934dc364 not found: ID does not exist" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.446593 4805 scope.go:117] "RemoveContainer" containerID="ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76" Nov 28 15:47:14 crc kubenswrapper[4805]: E1128 15:47:14.446967 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76\": container with ID starting with ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76 not found: ID does not exist" containerID="ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.447016 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76"} err="failed to get container status \"ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76\": rpc error: code = NotFound desc = could not find container \"ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76\": container with ID starting with ff4beec375af3201c98ea5378d381b9e3a14e1775ef546ad44a0043cd68a0f76 not found: ID does not exist" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.447049 4805 scope.go:117] "RemoveContainer" containerID="ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974" Nov 28 15:47:14 crc kubenswrapper[4805]: E1128 15:47:14.447468 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974\": container with ID starting with ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974 not found: ID does not exist" containerID="ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.447500 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974"} err="failed to get container status \"ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974\": rpc error: code = NotFound desc = could not find container \"ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974\": container with ID starting with ddfa121f585a3cc9244cc7f7163588074c9c330ec138e27f41ff815f3b04c974 not found: ID does not exist" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.447521 4805 scope.go:117] "RemoveContainer" containerID="d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579" Nov 28 15:47:14 crc kubenswrapper[4805]: E1128 15:47:14.447887 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579\": container with ID starting with d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579 not found: ID does not exist" containerID="d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.447913 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579"} err="failed to get container status \"d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579\": rpc error: code = NotFound desc = could not find container \"d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579\": container with ID starting with d7505476360ba36662b19b4ecf49af66ee74f54c760d52cde7aee8be38e4f579 not found: ID does not exist" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.510523 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.510501057 podStartE2EDuration="3.510501057s" podCreationTimestamp="2025-11-28 15:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:14.500008881 +0000 UTC m=+1261.549800192" watchObservedRunningTime="2025-11-28 15:47:14.510501057 +0000 UTC m=+1261.560292368" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.548385 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-sg-core-conf-yaml\") pod \"7ade9363-30e9-45e1-a57b-00704d8277ca\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.548473 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-run-httpd\") pod \"7ade9363-30e9-45e1-a57b-00704d8277ca\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.548503 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-config-data\") pod \"7ade9363-30e9-45e1-a57b-00704d8277ca\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.548572 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-scripts\") pod \"7ade9363-30e9-45e1-a57b-00704d8277ca\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.548655 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-combined-ca-bundle\") pod \"7ade9363-30e9-45e1-a57b-00704d8277ca\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.548736 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-log-httpd\") pod \"7ade9363-30e9-45e1-a57b-00704d8277ca\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.548815 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2szs\" (UniqueName: \"kubernetes.io/projected/7ade9363-30e9-45e1-a57b-00704d8277ca-kube-api-access-l2szs\") pod \"7ade9363-30e9-45e1-a57b-00704d8277ca\" (UID: \"7ade9363-30e9-45e1-a57b-00704d8277ca\") " Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.549410 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7ade9363-30e9-45e1-a57b-00704d8277ca" (UID: "7ade9363-30e9-45e1-a57b-00704d8277ca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.549803 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7ade9363-30e9-45e1-a57b-00704d8277ca" (UID: "7ade9363-30e9-45e1-a57b-00704d8277ca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.555554 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-scripts" (OuterVolumeSpecName: "scripts") pod "7ade9363-30e9-45e1-a57b-00704d8277ca" (UID: "7ade9363-30e9-45e1-a57b-00704d8277ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.555646 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ade9363-30e9-45e1-a57b-00704d8277ca-kube-api-access-l2szs" (OuterVolumeSpecName: "kube-api-access-l2szs") pod "7ade9363-30e9-45e1-a57b-00704d8277ca" (UID: "7ade9363-30e9-45e1-a57b-00704d8277ca"). InnerVolumeSpecName "kube-api-access-l2szs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.579385 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7ade9363-30e9-45e1-a57b-00704d8277ca" (UID: "7ade9363-30e9-45e1-a57b-00704d8277ca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.651241 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.651274 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2szs\" (UniqueName: \"kubernetes.io/projected/7ade9363-30e9-45e1-a57b-00704d8277ca-kube-api-access-l2szs\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.651284 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.651293 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ade9363-30e9-45e1-a57b-00704d8277ca-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.651301 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.661349 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ade9363-30e9-45e1-a57b-00704d8277ca" (UID: "7ade9363-30e9-45e1-a57b-00704d8277ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.691112 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-config-data" (OuterVolumeSpecName: "config-data") pod "7ade9363-30e9-45e1-a57b-00704d8277ca" (UID: "7ade9363-30e9-45e1-a57b-00704d8277ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.753200 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.753241 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ade9363-30e9-45e1-a57b-00704d8277ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.971739 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:14 crc kubenswrapper[4805]: I1128 15:47:14.979684 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.015530 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:15 crc kubenswrapper[4805]: E1128 15:47:15.015949 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="ceilometer-notification-agent" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.015967 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="ceilometer-notification-agent" Nov 28 15:47:15 crc kubenswrapper[4805]: E1128 15:47:15.015982 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="sg-core" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.016004 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="sg-core" Nov 28 15:47:15 crc kubenswrapper[4805]: E1128 15:47:15.016017 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="ceilometer-central-agent" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.016023 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="ceilometer-central-agent" Nov 28 15:47:15 crc kubenswrapper[4805]: E1128 15:47:15.016049 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="proxy-httpd" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.016056 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="proxy-httpd" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.016218 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="ceilometer-notification-agent" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.016229 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="ceilometer-central-agent" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.016242 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="proxy-httpd" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.016258 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" containerName="sg-core" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.017755 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.020255 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.020499 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.033553 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.168815 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8whh9\" (UniqueName: \"kubernetes.io/projected/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-kube-api-access-8whh9\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.168904 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-run-httpd\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.168967 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-config-data\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.169031 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.169053 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-scripts\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.169089 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.169117 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-log-httpd\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.218402 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ade9363-30e9-45e1-a57b-00704d8277ca" path="/var/lib/kubelet/pods/7ade9363-30e9-45e1-a57b-00704d8277ca/volumes" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.270545 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-run-httpd\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.270606 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-config-data\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.270664 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.270686 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-scripts\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.270719 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.270745 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-log-httpd\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.270917 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8whh9\" (UniqueName: \"kubernetes.io/projected/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-kube-api-access-8whh9\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.271920 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-run-httpd\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.272260 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-log-httpd\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.276285 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.276775 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.277070 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-scripts\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.278347 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-config-data\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.300163 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8whh9\" (UniqueName: \"kubernetes.io/projected/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-kube-api-access-8whh9\") pod \"ceilometer-0\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.338635 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.338663 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.348518 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.707034 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:15 crc kubenswrapper[4805]: I1128 15:47:15.824460 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 15:47:18 crc kubenswrapper[4805]: I1128 15:47:18.888650 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:21 crc kubenswrapper[4805]: I1128 15:47:21.064658 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:21 crc kubenswrapper[4805]: W1128 15:47:21.069990 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e2d5f1d_ee6a_42fe_acd6_b64cb9d77932.slice/crio-f887ae09389ed89b5e26616d8ec9d24072425dbe2c380b5e6690c5e9a2b6aa56 WatchSource:0}: Error finding container f887ae09389ed89b5e26616d8ec9d24072425dbe2c380b5e6690c5e9a2b6aa56: Status 404 returned error can't find the container with id f887ae09389ed89b5e26616d8ec9d24072425dbe2c380b5e6690c5e9a2b6aa56 Nov 28 15:47:21 crc kubenswrapper[4805]: I1128 15:47:21.413679 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerStarted","Data":"f887ae09389ed89b5e26616d8ec9d24072425dbe2c380b5e6690c5e9a2b6aa56"} Nov 28 15:47:21 crc kubenswrapper[4805]: I1128 15:47:21.415666 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cfqg8" event={"ID":"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6","Type":"ContainerStarted","Data":"b6992038b195926e6c2ae3d4f1f5dd8116283fe079e24263ee47fb35fb22332a"} Nov 28 15:47:21 crc kubenswrapper[4805]: I1128 15:47:21.434835 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-cfqg8" podStartSLOduration=1.726183903 podStartE2EDuration="11.434820879s" podCreationTimestamp="2025-11-28 15:47:10 +0000 UTC" firstStartedPulling="2025-11-28 15:47:10.943520098 +0000 UTC m=+1257.993311419" lastFinishedPulling="2025-11-28 15:47:20.652157084 +0000 UTC m=+1267.701948395" observedRunningTime="2025-11-28 15:47:21.432575278 +0000 UTC m=+1268.482366599" watchObservedRunningTime="2025-11-28 15:47:21.434820879 +0000 UTC m=+1268.484612190" Nov 28 15:47:21 crc kubenswrapper[4805]: I1128 15:47:21.741390 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 15:47:21 crc kubenswrapper[4805]: I1128 15:47:21.741472 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 15:47:21 crc kubenswrapper[4805]: I1128 15:47:21.780068 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 15:47:21 crc kubenswrapper[4805]: I1128 15:47:21.789155 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 15:47:22 crc kubenswrapper[4805]: I1128 15:47:22.424400 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerStarted","Data":"3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b"} Nov 28 15:47:22 crc kubenswrapper[4805]: I1128 15:47:22.424778 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 15:47:22 crc kubenswrapper[4805]: I1128 15:47:22.424796 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 15:47:23 crc kubenswrapper[4805]: I1128 15:47:23.437124 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerStarted","Data":"60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64"} Nov 28 15:47:24 crc kubenswrapper[4805]: I1128 15:47:24.448044 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerStarted","Data":"702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9"} Nov 28 15:47:24 crc kubenswrapper[4805]: I1128 15:47:24.502164 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 15:47:24 crc kubenswrapper[4805]: I1128 15:47:24.502267 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 15:47:24 crc kubenswrapper[4805]: I1128 15:47:24.585170 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 15:47:29 crc kubenswrapper[4805]: I1128 15:47:29.496840 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerStarted","Data":"0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93"} Nov 28 15:47:29 crc kubenswrapper[4805]: I1128 15:47:29.498464 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 15:47:29 crc kubenswrapper[4805]: I1128 15:47:29.497733 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="proxy-httpd" containerID="cri-o://0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93" gracePeriod=30 Nov 28 15:47:29 crc kubenswrapper[4805]: I1128 15:47:29.497133 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="ceilometer-central-agent" containerID="cri-o://3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b" gracePeriod=30 Nov 28 15:47:29 crc kubenswrapper[4805]: I1128 15:47:29.497747 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="sg-core" containerID="cri-o://702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9" gracePeriod=30 Nov 28 15:47:29 crc kubenswrapper[4805]: I1128 15:47:29.497756 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="ceilometer-notification-agent" containerID="cri-o://60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64" gracePeriod=30 Nov 28 15:47:29 crc kubenswrapper[4805]: I1128 15:47:29.522118 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=7.882109255 podStartE2EDuration="15.522075788s" podCreationTimestamp="2025-11-28 15:47:14 +0000 UTC" firstStartedPulling="2025-11-28 15:47:21.072826955 +0000 UTC m=+1268.122618266" lastFinishedPulling="2025-11-28 15:47:28.712793478 +0000 UTC m=+1275.762584799" observedRunningTime="2025-11-28 15:47:29.518788699 +0000 UTC m=+1276.568580020" watchObservedRunningTime="2025-11-28 15:47:29.522075788 +0000 UTC m=+1276.571867099" Nov 28 15:47:30 crc kubenswrapper[4805]: I1128 15:47:30.528792 4805 generic.go:334] "Generic (PLEG): container finished" podID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerID="0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93" exitCode=0 Nov 28 15:47:30 crc kubenswrapper[4805]: I1128 15:47:30.528870 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerDied","Data":"0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93"} Nov 28 15:47:30 crc kubenswrapper[4805]: I1128 15:47:30.528910 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerDied","Data":"702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9"} Nov 28 15:47:30 crc kubenswrapper[4805]: I1128 15:47:30.529028 4805 generic.go:334] "Generic (PLEG): container finished" podID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerID="702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9" exitCode=2 Nov 28 15:47:30 crc kubenswrapper[4805]: I1128 15:47:30.529101 4805 generic.go:334] "Generic (PLEG): container finished" podID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerID="60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64" exitCode=0 Nov 28 15:47:30 crc kubenswrapper[4805]: I1128 15:47:30.529135 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerDied","Data":"60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64"} Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.081559 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.152606 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-run-httpd\") pod \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.153658 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-scripts\") pod \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.153796 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-combined-ca-bundle\") pod \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.153993 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-log-httpd\") pod \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.154713 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-sg-core-conf-yaml\") pod \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.154877 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-config-data\") pod \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.154946 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8whh9\" (UniqueName: \"kubernetes.io/projected/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-kube-api-access-8whh9\") pod \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\" (UID: \"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932\") " Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.156037 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" (UID: "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.157329 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" (UID: "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.157955 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.157981 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.161673 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-scripts" (OuterVolumeSpecName: "scripts") pod "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" (UID: "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.162596 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-kube-api-access-8whh9" (OuterVolumeSpecName: "kube-api-access-8whh9") pod "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" (UID: "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932"). InnerVolumeSpecName "kube-api-access-8whh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.192081 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" (UID: "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.259522 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.259549 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8whh9\" (UniqueName: \"kubernetes.io/projected/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-kube-api-access-8whh9\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.259557 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.262808 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" (UID: "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.272625 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-config-data" (OuterVolumeSpecName: "config-data") pod "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" (UID: "5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.361865 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.361948 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.573096 4805 generic.go:334] "Generic (PLEG): container finished" podID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerID="3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b" exitCode=0 Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.573199 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.573315 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerDied","Data":"3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b"} Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.573414 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932","Type":"ContainerDied","Data":"f887ae09389ed89b5e26616d8ec9d24072425dbe2c380b5e6690c5e9a2b6aa56"} Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.573456 4805 scope.go:117] "RemoveContainer" containerID="0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.600952 4805 scope.go:117] "RemoveContainer" containerID="702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.627028 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.638689 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.644529 4805 scope.go:117] "RemoveContainer" containerID="60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.653893 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:34 crc kubenswrapper[4805]: E1128 15:47:34.654444 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="proxy-httpd" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.654461 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="proxy-httpd" Nov 28 15:47:34 crc kubenswrapper[4805]: E1128 15:47:34.654484 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="ceilometer-notification-agent" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.654490 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="ceilometer-notification-agent" Nov 28 15:47:34 crc kubenswrapper[4805]: E1128 15:47:34.654505 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="ceilometer-central-agent" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.654513 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="ceilometer-central-agent" Nov 28 15:47:34 crc kubenswrapper[4805]: E1128 15:47:34.654534 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="sg-core" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.654540 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="sg-core" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.654774 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="proxy-httpd" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.654791 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="ceilometer-central-agent" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.654799 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="ceilometer-notification-agent" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.654817 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" containerName="sg-core" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.656809 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.663549 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.663853 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.666902 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgwdb\" (UniqueName: \"kubernetes.io/projected/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-kube-api-access-dgwdb\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.666973 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.667060 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-scripts\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.667097 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-log-httpd\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.667140 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-config-data\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.667202 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.667246 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-run-httpd\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.674169 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.689405 4805 scope.go:117] "RemoveContainer" containerID="3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.713689 4805 scope.go:117] "RemoveContainer" containerID="0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93" Nov 28 15:47:34 crc kubenswrapper[4805]: E1128 15:47:34.714170 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93\": container with ID starting with 0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93 not found: ID does not exist" containerID="0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.714208 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93"} err="failed to get container status \"0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93\": rpc error: code = NotFound desc = could not find container \"0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93\": container with ID starting with 0bab75ef3eef8d1e48b46bad759cc89ee1c79ff8851a2313302b3dd6868c9c93 not found: ID does not exist" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.714238 4805 scope.go:117] "RemoveContainer" containerID="702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9" Nov 28 15:47:34 crc kubenswrapper[4805]: E1128 15:47:34.714705 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9\": container with ID starting with 702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9 not found: ID does not exist" containerID="702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.714740 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9"} err="failed to get container status \"702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9\": rpc error: code = NotFound desc = could not find container \"702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9\": container with ID starting with 702c1ec35c87964d9c44e277332a4e36d84b7dec1ee0af46134188be1a4329e9 not found: ID does not exist" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.714760 4805 scope.go:117] "RemoveContainer" containerID="60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64" Nov 28 15:47:34 crc kubenswrapper[4805]: E1128 15:47:34.715118 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64\": container with ID starting with 60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64 not found: ID does not exist" containerID="60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.715145 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64"} err="failed to get container status \"60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64\": rpc error: code = NotFound desc = could not find container \"60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64\": container with ID starting with 60ab875ac07b451fc90934f5bae004f26eba97164da9d8b54d6c6440a9e9ef64 not found: ID does not exist" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.715161 4805 scope.go:117] "RemoveContainer" containerID="3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b" Nov 28 15:47:34 crc kubenswrapper[4805]: E1128 15:47:34.715443 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b\": container with ID starting with 3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b not found: ID does not exist" containerID="3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.715474 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b"} err="failed to get container status \"3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b\": rpc error: code = NotFound desc = could not find container \"3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b\": container with ID starting with 3a72933da95f40eea180ac8b127e128ca02506ac7eb6af742a616a2c2808705b not found: ID does not exist" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.768709 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-config-data\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.768786 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.768826 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-run-httpd\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.768860 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgwdb\" (UniqueName: \"kubernetes.io/projected/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-kube-api-access-dgwdb\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.768898 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.768957 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-scripts\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.768988 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-log-httpd\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.769537 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-log-httpd\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.772477 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-run-httpd\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.773488 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-config-data\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.775091 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-scripts\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.776007 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.781073 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.784731 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgwdb\" (UniqueName: \"kubernetes.io/projected/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-kube-api-access-dgwdb\") pod \"ceilometer-0\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " pod="openstack/ceilometer-0" Nov 28 15:47:34 crc kubenswrapper[4805]: I1128 15:47:34.979544 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:47:35 crc kubenswrapper[4805]: I1128 15:47:35.216630 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932" path="/var/lib/kubelet/pods/5e2d5f1d-ee6a-42fe-acd6-b64cb9d77932/volumes" Nov 28 15:47:35 crc kubenswrapper[4805]: I1128 15:47:35.508637 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:47:35 crc kubenswrapper[4805]: W1128 15:47:35.512670 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7cb2c0c_ecfc_4eda_9d06_25e568535d43.slice/crio-1ed60d0035061e1aeb8e2238fdd14e3701a46141a1b732980041fa08471f8efb WatchSource:0}: Error finding container 1ed60d0035061e1aeb8e2238fdd14e3701a46141a1b732980041fa08471f8efb: Status 404 returned error can't find the container with id 1ed60d0035061e1aeb8e2238fdd14e3701a46141a1b732980041fa08471f8efb Nov 28 15:47:35 crc kubenswrapper[4805]: I1128 15:47:35.586169 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerStarted","Data":"1ed60d0035061e1aeb8e2238fdd14e3701a46141a1b732980041fa08471f8efb"} Nov 28 15:47:37 crc kubenswrapper[4805]: I1128 15:47:37.637937 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerStarted","Data":"76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f"} Nov 28 15:47:38 crc kubenswrapper[4805]: I1128 15:47:38.646978 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerStarted","Data":"51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3"} Nov 28 15:47:39 crc kubenswrapper[4805]: I1128 15:47:39.676952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerStarted","Data":"231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7"} Nov 28 15:47:41 crc kubenswrapper[4805]: I1128 15:47:41.060640 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:47:41 crc kubenswrapper[4805]: I1128 15:47:41.061375 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:47:41 crc kubenswrapper[4805]: I1128 15:47:41.700406 4805 generic.go:334] "Generic (PLEG): container finished" podID="547a6f46-a3b2-42b9-9681-80e8bbfaa4e6" containerID="b6992038b195926e6c2ae3d4f1f5dd8116283fe079e24263ee47fb35fb22332a" exitCode=0 Nov 28 15:47:41 crc kubenswrapper[4805]: I1128 15:47:41.700572 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cfqg8" event={"ID":"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6","Type":"ContainerDied","Data":"b6992038b195926e6c2ae3d4f1f5dd8116283fe079e24263ee47fb35fb22332a"} Nov 28 15:47:41 crc kubenswrapper[4805]: I1128 15:47:41.703751 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerStarted","Data":"6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e"} Nov 28 15:47:41 crc kubenswrapper[4805]: I1128 15:47:41.704493 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 15:47:41 crc kubenswrapper[4805]: I1128 15:47:41.744837 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.58526037 podStartE2EDuration="7.744819032s" podCreationTimestamp="2025-11-28 15:47:34 +0000 UTC" firstStartedPulling="2025-11-28 15:47:35.515469068 +0000 UTC m=+1282.565260379" lastFinishedPulling="2025-11-28 15:47:40.67502771 +0000 UTC m=+1287.724819041" observedRunningTime="2025-11-28 15:47:41.732887728 +0000 UTC m=+1288.782679059" watchObservedRunningTime="2025-11-28 15:47:41.744819032 +0000 UTC m=+1288.794610343" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.151147 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.227184 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-config-data\") pod \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.227692 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9fh6\" (UniqueName: \"kubernetes.io/projected/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-kube-api-access-r9fh6\") pod \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.227863 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-scripts\") pod \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.228007 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-combined-ca-bundle\") pod \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\" (UID: \"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6\") " Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.232992 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-kube-api-access-r9fh6" (OuterVolumeSpecName: "kube-api-access-r9fh6") pod "547a6f46-a3b2-42b9-9681-80e8bbfaa4e6" (UID: "547a6f46-a3b2-42b9-9681-80e8bbfaa4e6"). InnerVolumeSpecName "kube-api-access-r9fh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.234325 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-scripts" (OuterVolumeSpecName: "scripts") pod "547a6f46-a3b2-42b9-9681-80e8bbfaa4e6" (UID: "547a6f46-a3b2-42b9-9681-80e8bbfaa4e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.274724 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "547a6f46-a3b2-42b9-9681-80e8bbfaa4e6" (UID: "547a6f46-a3b2-42b9-9681-80e8bbfaa4e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.280022 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-config-data" (OuterVolumeSpecName: "config-data") pod "547a6f46-a3b2-42b9-9681-80e8bbfaa4e6" (UID: "547a6f46-a3b2-42b9-9681-80e8bbfaa4e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.330422 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.330457 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.330591 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.330859 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9fh6\" (UniqueName: \"kubernetes.io/projected/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6-kube-api-access-r9fh6\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.725883 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cfqg8" event={"ID":"547a6f46-a3b2-42b9-9681-80e8bbfaa4e6","Type":"ContainerDied","Data":"26625d56ecaa27a9ae9e5f970565463dd5d522366994d0b1f08be2e4a28548f5"} Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.725904 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cfqg8" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.725921 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26625d56ecaa27a9ae9e5f970565463dd5d522366994d0b1f08be2e4a28548f5" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.825137 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 15:47:43 crc kubenswrapper[4805]: E1128 15:47:43.825553 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547a6f46-a3b2-42b9-9681-80e8bbfaa4e6" containerName="nova-cell0-conductor-db-sync" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.825569 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="547a6f46-a3b2-42b9-9681-80e8bbfaa4e6" containerName="nova-cell0-conductor-db-sync" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.825769 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="547a6f46-a3b2-42b9-9681-80e8bbfaa4e6" containerName="nova-cell0-conductor-db-sync" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.826323 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.828680 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.828847 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-7gz9m" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.842616 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.946152 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7nvm\" (UniqueName: \"kubernetes.io/projected/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-kube-api-access-g7nvm\") pod \"nova-cell0-conductor-0\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.946214 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:43 crc kubenswrapper[4805]: I1128 15:47:43.946317 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:44 crc kubenswrapper[4805]: I1128 15:47:44.047911 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7nvm\" (UniqueName: \"kubernetes.io/projected/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-kube-api-access-g7nvm\") pod \"nova-cell0-conductor-0\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:44 crc kubenswrapper[4805]: I1128 15:47:44.047996 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:44 crc kubenswrapper[4805]: I1128 15:47:44.048133 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:44 crc kubenswrapper[4805]: I1128 15:47:44.053582 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:44 crc kubenswrapper[4805]: I1128 15:47:44.062247 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:44 crc kubenswrapper[4805]: I1128 15:47:44.068049 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7nvm\" (UniqueName: \"kubernetes.io/projected/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-kube-api-access-g7nvm\") pod \"nova-cell0-conductor-0\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:44 crc kubenswrapper[4805]: I1128 15:47:44.160006 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:44 crc kubenswrapper[4805]: I1128 15:47:44.612884 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 15:47:44 crc kubenswrapper[4805]: I1128 15:47:44.733547 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1","Type":"ContainerStarted","Data":"82fb7f71741c2e2edd378bc77ad4fdff334b437d92ae7263f66e56dccca3ee9e"} Nov 28 15:47:45 crc kubenswrapper[4805]: I1128 15:47:45.743864 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1","Type":"ContainerStarted","Data":"3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339"} Nov 28 15:47:45 crc kubenswrapper[4805]: I1128 15:47:45.745291 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:45 crc kubenswrapper[4805]: I1128 15:47:45.765716 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.7656966069999998 podStartE2EDuration="2.765696607s" podCreationTimestamp="2025-11-28 15:47:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:45.760820605 +0000 UTC m=+1292.810611926" watchObservedRunningTime="2025-11-28 15:47:45.765696607 +0000 UTC m=+1292.815487928" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.187510 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.616901 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-bllqv"] Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.618546 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.622022 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.624613 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.647772 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bllqv"] Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.755839 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-config-data\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.755942 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-scripts\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.756044 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.756099 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxnj9\" (UniqueName: \"kubernetes.io/projected/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-kube-api-access-lxnj9\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.765019 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.766311 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.768427 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.782271 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.857910 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.858001 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxnj9\" (UniqueName: \"kubernetes.io/projected/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-kube-api-access-lxnj9\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.858050 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " pod="openstack/nova-scheduler-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.858097 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-config-data\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.858125 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp4s2\" (UniqueName: \"kubernetes.io/projected/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-kube-api-access-jp4s2\") pod \"nova-scheduler-0\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " pod="openstack/nova-scheduler-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.858191 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-scripts\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.858265 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-config-data\") pod \"nova-scheduler-0\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " pod="openstack/nova-scheduler-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.867256 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-scripts\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.868002 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-config-data\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.868712 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.891412 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxnj9\" (UniqueName: \"kubernetes.io/projected/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-kube-api-access-lxnj9\") pod \"nova-cell0-cell-mapping-bllqv\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.910532 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.912408 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.916688 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.938420 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.956892 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.958794 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.967595 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-config-data\") pod \"nova-scheduler-0\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " pod="openstack/nova-scheduler-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.968033 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.970894 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " pod="openstack/nova-scheduler-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.971031 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp4s2\" (UniqueName: \"kubernetes.io/projected/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-kube-api-access-jp4s2\") pod \"nova-scheduler-0\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " pod="openstack/nova-scheduler-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.986975 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-config-data\") pod \"nova-scheduler-0\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " pod="openstack/nova-scheduler-0" Nov 28 15:47:49 crc kubenswrapper[4805]: I1128 15:47:49.997654 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " pod="openstack/nova-scheduler-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.014194 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.014808 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp4s2\" (UniqueName: \"kubernetes.io/projected/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-kube-api-access-jp4s2\") pod \"nova-scheduler-0\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " pod="openstack/nova-scheduler-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.037413 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.072846 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.074501 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.072695 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fc90a98-5870-4ba2-8c31-0891019c9ac4-logs\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.076261 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.076566 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqfqn\" (UniqueName: \"kubernetes.io/projected/b7739baa-2c83-453a-9bb7-539cff74b9f7-kube-api-access-sqfqn\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.076745 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7739baa-2c83-453a-9bb7-539cff74b9f7-logs\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.076877 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-config-data\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.076975 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf27b\" (UniqueName: \"kubernetes.io/projected/0fc90a98-5870-4ba2-8c31-0891019c9ac4-kube-api-access-pf27b\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.077151 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.076573 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.077487 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-config-data\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.091825 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.134636 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-8ln5n"] Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.136201 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.150433 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.180414 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7739baa-2c83-453a-9bb7-539cff74b9f7-logs\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.180590 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf27b\" (UniqueName: \"kubernetes.io/projected/0fc90a98-5870-4ba2-8c31-0891019c9ac4-kube-api-access-pf27b\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.180629 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-config-data\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.180709 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.180800 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-config-data\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.180932 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.180976 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7739baa-2c83-453a-9bb7-539cff74b9f7-logs\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.181013 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.181104 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vtrz\" (UniqueName: \"kubernetes.io/projected/7d2addf2-9fd9-4f75-804c-7668ca2684a3-kube-api-access-9vtrz\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.181159 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fc90a98-5870-4ba2-8c31-0891019c9ac4-logs\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.181182 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.181236 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqfqn\" (UniqueName: \"kubernetes.io/projected/b7739baa-2c83-453a-9bb7-539cff74b9f7-kube-api-access-sqfqn\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.183508 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-8ln5n"] Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.184010 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fc90a98-5870-4ba2-8c31-0891019c9ac4-logs\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.191806 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.191845 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-config-data\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.200873 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.204050 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-config-data\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.204577 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqfqn\" (UniqueName: \"kubernetes.io/projected/b7739baa-2c83-453a-9bb7-539cff74b9f7-kube-api-access-sqfqn\") pod \"nova-api-0\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.208294 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf27b\" (UniqueName: \"kubernetes.io/projected/0fc90a98-5870-4ba2-8c31-0891019c9ac4-kube-api-access-pf27b\") pod \"nova-metadata-0\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.282754 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6zjh\" (UniqueName: \"kubernetes.io/projected/09e077c7-c008-4dae-af71-45fc22923773-kube-api-access-g6zjh\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.283132 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vtrz\" (UniqueName: \"kubernetes.io/projected/7d2addf2-9fd9-4f75-804c-7668ca2684a3-kube-api-access-9vtrz\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.283525 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-config\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.283555 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-swift-storage-0\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.283626 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-sb\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.283655 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-svc\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.283760 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-nb\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.289708 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.289807 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.293868 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.295280 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.307967 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vtrz\" (UniqueName: \"kubernetes.io/projected/7d2addf2-9fd9-4f75-804c-7668ca2684a3-kube-api-access-9vtrz\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.391712 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-nb\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.391861 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6zjh\" (UniqueName: \"kubernetes.io/projected/09e077c7-c008-4dae-af71-45fc22923773-kube-api-access-g6zjh\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.391935 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-config\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.391957 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-swift-storage-0\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.391993 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-sb\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.392016 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-svc\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.393445 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-swift-storage-0\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.393671 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-svc\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.393673 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-sb\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.394258 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-nb\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.395491 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-config\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.415381 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.426781 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.432044 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6zjh\" (UniqueName: \"kubernetes.io/projected/09e077c7-c008-4dae-af71-45fc22923773-kube-api-access-g6zjh\") pod \"dnsmasq-dns-5594d9b959-8ln5n\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.448752 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.474034 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.610184 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.631601 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bllqv"] Nov 28 15:47:50 crc kubenswrapper[4805]: W1128 15:47:50.654953 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6d86dfa_8c46_40bb_8155_438f6d5e06e6.slice/crio-521f2d671001675d6624cd237bde3eb45a36d6137bcf47b7eba7184b9cc3b29c WatchSource:0}: Error finding container 521f2d671001675d6624cd237bde3eb45a36d6137bcf47b7eba7184b9cc3b29c: Status 404 returned error can't find the container with id 521f2d671001675d6624cd237bde3eb45a36d6137bcf47b7eba7184b9cc3b29c Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.804441 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb","Type":"ContainerStarted","Data":"f516e2b11362eac8f0cf8b0be9c17f460c3bdf65050d542c4e738721a117392c"} Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.807148 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bllqv" event={"ID":"b6d86dfa-8c46-40bb-8155-438f6d5e06e6","Type":"ContainerStarted","Data":"521f2d671001675d6624cd237bde3eb45a36d6137bcf47b7eba7184b9cc3b29c"} Nov 28 15:47:50 crc kubenswrapper[4805]: I1128 15:47:50.972465 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:47:50 crc kubenswrapper[4805]: W1128 15:47:50.978404 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fc90a98_5870_4ba2_8c31_0891019c9ac4.slice/crio-4a1e5de63d93c96b0da1c4a180d2725d843d4b971ac38fc773b5cdaca4a4dbb6 WatchSource:0}: Error finding container 4a1e5de63d93c96b0da1c4a180d2725d843d4b971ac38fc773b5cdaca4a4dbb6: Status 404 returned error can't find the container with id 4a1e5de63d93c96b0da1c4a180d2725d843d4b971ac38fc773b5cdaca4a4dbb6 Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.036884 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pc5z4"] Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.038021 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.042605 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.043644 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.050125 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pc5z4"] Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.136520 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.189733 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.245759 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-scripts\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.246346 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-config-data\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.246519 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.246812 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khjd5\" (UniqueName: \"kubernetes.io/projected/39c6f2be-4e1d-40c1-8871-a008f39d2d49-kube-api-access-khjd5\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.283662 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-8ln5n"] Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.369292 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-scripts\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.369406 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-config-data\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.369452 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.369536 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khjd5\" (UniqueName: \"kubernetes.io/projected/39c6f2be-4e1d-40c1-8871-a008f39d2d49-kube-api-access-khjd5\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.374662 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.374695 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-config-data\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.375024 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-scripts\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.386229 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khjd5\" (UniqueName: \"kubernetes.io/projected/39c6f2be-4e1d-40c1-8871-a008f39d2d49-kube-api-access-khjd5\") pod \"nova-cell1-conductor-db-sync-pc5z4\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.662988 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.820003 4805 generic.go:334] "Generic (PLEG): container finished" podID="09e077c7-c008-4dae-af71-45fc22923773" containerID="a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269" exitCode=0 Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.820066 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" event={"ID":"09e077c7-c008-4dae-af71-45fc22923773","Type":"ContainerDied","Data":"a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269"} Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.820092 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" event={"ID":"09e077c7-c008-4dae-af71-45fc22923773","Type":"ContainerStarted","Data":"77e9d95935a04d8ed847507b1a68facebee44c7fdbfd574578a6027db33e65a5"} Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.841286 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b7739baa-2c83-453a-9bb7-539cff74b9f7","Type":"ContainerStarted","Data":"3161af124e6754f6db4d0d8859e685ae207735e0861f1754fdfcfdb3ea662c99"} Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.850164 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bllqv" event={"ID":"b6d86dfa-8c46-40bb-8155-438f6d5e06e6","Type":"ContainerStarted","Data":"c8dee08d7ace9297ae12c4f0d44cad8e080785d18e73355b29ebf5cc4b33729a"} Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.864452 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7d2addf2-9fd9-4f75-804c-7668ca2684a3","Type":"ContainerStarted","Data":"7507b27a704e4c73b66fae6f62e63c2f8cc70f46c910935cb6bb3659dc88c867"} Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.868110 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fc90a98-5870-4ba2-8c31-0891019c9ac4","Type":"ContainerStarted","Data":"4a1e5de63d93c96b0da1c4a180d2725d843d4b971ac38fc773b5cdaca4a4dbb6"} Nov 28 15:47:51 crc kubenswrapper[4805]: I1128 15:47:51.877287 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-bllqv" podStartSLOduration=2.877271807 podStartE2EDuration="2.877271807s" podCreationTimestamp="2025-11-28 15:47:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:51.876180696 +0000 UTC m=+1298.925972007" watchObservedRunningTime="2025-11-28 15:47:51.877271807 +0000 UTC m=+1298.927063118" Nov 28 15:47:52 crc kubenswrapper[4805]: I1128 15:47:52.751919 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pc5z4"] Nov 28 15:47:52 crc kubenswrapper[4805]: I1128 15:47:52.883126 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" event={"ID":"09e077c7-c008-4dae-af71-45fc22923773","Type":"ContainerStarted","Data":"968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06"} Nov 28 15:47:52 crc kubenswrapper[4805]: I1128 15:47:52.884559 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:47:52 crc kubenswrapper[4805]: I1128 15:47:52.888043 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb","Type":"ContainerStarted","Data":"0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8"} Nov 28 15:47:52 crc kubenswrapper[4805]: I1128 15:47:52.904261 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" podStartSLOduration=3.904243292 podStartE2EDuration="3.904243292s" podCreationTimestamp="2025-11-28 15:47:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:52.900178752 +0000 UTC m=+1299.949970063" watchObservedRunningTime="2025-11-28 15:47:52.904243292 +0000 UTC m=+1299.954034603" Nov 28 15:47:52 crc kubenswrapper[4805]: I1128 15:47:52.920250 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.265521053 podStartE2EDuration="3.920234468s" podCreationTimestamp="2025-11-28 15:47:49 +0000 UTC" firstStartedPulling="2025-11-28 15:47:50.64787987 +0000 UTC m=+1297.697671181" lastFinishedPulling="2025-11-28 15:47:52.302593285 +0000 UTC m=+1299.352384596" observedRunningTime="2025-11-28 15:47:52.917533824 +0000 UTC m=+1299.967325145" watchObservedRunningTime="2025-11-28 15:47:52.920234468 +0000 UTC m=+1299.970025779" Nov 28 15:47:53 crc kubenswrapper[4805]: I1128 15:47:53.901554 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pc5z4" event={"ID":"39c6f2be-4e1d-40c1-8871-a008f39d2d49","Type":"ContainerStarted","Data":"03396b519d892da63f88c2cb02e47351ec96ea9afc54700a2a63f380098045fd"} Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.270110 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.303415 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.912490 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b7739baa-2c83-453a-9bb7-539cff74b9f7","Type":"ContainerStarted","Data":"ba50f2ffcc3947dbd305c2c3dbf7b3bea9c39237b598f372706ab0a19f01827c"} Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.912561 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b7739baa-2c83-453a-9bb7-539cff74b9f7","Type":"ContainerStarted","Data":"358443cbcb7f69cbca8e140908a93dcf6c47bd8d6c4023ba06d915c0a12242d9"} Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.915591 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7d2addf2-9fd9-4f75-804c-7668ca2684a3","Type":"ContainerStarted","Data":"dfeb9cf2753f320f8bad6ffc3d93be1753051da92f071ac9cfa6ec09ca20c114"} Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.915707 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="7d2addf2-9fd9-4f75-804c-7668ca2684a3" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://dfeb9cf2753f320f8bad6ffc3d93be1753051da92f071ac9cfa6ec09ca20c114" gracePeriod=30 Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.924914 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fc90a98-5870-4ba2-8c31-0891019c9ac4","Type":"ContainerStarted","Data":"cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a"} Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.925201 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fc90a98-5870-4ba2-8c31-0891019c9ac4","Type":"ContainerStarted","Data":"0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f"} Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.925440 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerName="nova-metadata-log" containerID="cri-o://0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f" gracePeriod=30 Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.925613 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerName="nova-metadata-metadata" containerID="cri-o://cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a" gracePeriod=30 Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.927745 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pc5z4" event={"ID":"39c6f2be-4e1d-40c1-8871-a008f39d2d49","Type":"ContainerStarted","Data":"93594d7686a712e66ead92d80c77d078b1f2f43f80e381c3d84546645e2f43fc"} Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.959075 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.25684274 podStartE2EDuration="5.959050858s" podCreationTimestamp="2025-11-28 15:47:49 +0000 UTC" firstStartedPulling="2025-11-28 15:47:51.162576272 +0000 UTC m=+1298.212367583" lastFinishedPulling="2025-11-28 15:47:53.86478439 +0000 UTC m=+1300.914575701" observedRunningTime="2025-11-28 15:47:54.95248577 +0000 UTC m=+1302.002277101" watchObservedRunningTime="2025-11-28 15:47:54.959050858 +0000 UTC m=+1302.008842169" Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.960205 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.303060397 podStartE2EDuration="5.960198909s" podCreationTimestamp="2025-11-28 15:47:49 +0000 UTC" firstStartedPulling="2025-11-28 15:47:51.206329382 +0000 UTC m=+1298.256120683" lastFinishedPulling="2025-11-28 15:47:53.863467884 +0000 UTC m=+1300.913259195" observedRunningTime="2025-11-28 15:47:54.937600924 +0000 UTC m=+1301.987392245" watchObservedRunningTime="2025-11-28 15:47:54.960198909 +0000 UTC m=+1302.009990220" Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.977389 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-pc5z4" podStartSLOduration=3.977346826 podStartE2EDuration="3.977346826s" podCreationTimestamp="2025-11-28 15:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:54.966247854 +0000 UTC m=+1302.016039165" watchObservedRunningTime="2025-11-28 15:47:54.977346826 +0000 UTC m=+1302.027138137" Nov 28 15:47:54 crc kubenswrapper[4805]: I1128 15:47:54.999394 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.117429914 podStartE2EDuration="5.999352105s" podCreationTimestamp="2025-11-28 15:47:49 +0000 UTC" firstStartedPulling="2025-11-28 15:47:50.981173643 +0000 UTC m=+1298.030964954" lastFinishedPulling="2025-11-28 15:47:53.863095834 +0000 UTC m=+1300.912887145" observedRunningTime="2025-11-28 15:47:54.993050513 +0000 UTC m=+1302.042841844" watchObservedRunningTime="2025-11-28 15:47:54.999352105 +0000 UTC m=+1302.049143416" Nov 28 15:47:55 crc kubenswrapper[4805]: I1128 15:47:55.092950 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 15:47:55 crc kubenswrapper[4805]: I1128 15:47:55.416594 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 15:47:55 crc kubenswrapper[4805]: I1128 15:47:55.416656 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 15:47:55 crc kubenswrapper[4805]: I1128 15:47:55.448870 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:47:55 crc kubenswrapper[4805]: I1128 15:47:55.937455 4805 generic.go:334] "Generic (PLEG): container finished" podID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerID="0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f" exitCode=143 Nov 28 15:47:55 crc kubenswrapper[4805]: I1128 15:47:55.937484 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fc90a98-5870-4ba2-8c31-0891019c9ac4","Type":"ContainerDied","Data":"0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f"} Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.697739 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.812028 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fc90a98-5870-4ba2-8c31-0891019c9ac4-logs\") pod \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.812235 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-config-data\") pod \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.812389 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fc90a98-5870-4ba2-8c31-0891019c9ac4-logs" (OuterVolumeSpecName: "logs") pod "0fc90a98-5870-4ba2-8c31-0891019c9ac4" (UID: "0fc90a98-5870-4ba2-8c31-0891019c9ac4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.812459 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf27b\" (UniqueName: \"kubernetes.io/projected/0fc90a98-5870-4ba2-8c31-0891019c9ac4-kube-api-access-pf27b\") pod \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.812515 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-combined-ca-bundle\") pod \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\" (UID: \"0fc90a98-5870-4ba2-8c31-0891019c9ac4\") " Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.813411 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fc90a98-5870-4ba2-8c31-0891019c9ac4-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.820272 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fc90a98-5870-4ba2-8c31-0891019c9ac4-kube-api-access-pf27b" (OuterVolumeSpecName: "kube-api-access-pf27b") pod "0fc90a98-5870-4ba2-8c31-0891019c9ac4" (UID: "0fc90a98-5870-4ba2-8c31-0891019c9ac4"). InnerVolumeSpecName "kube-api-access-pf27b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.843938 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-config-data" (OuterVolumeSpecName: "config-data") pod "0fc90a98-5870-4ba2-8c31-0891019c9ac4" (UID: "0fc90a98-5870-4ba2-8c31-0891019c9ac4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.848964 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fc90a98-5870-4ba2-8c31-0891019c9ac4" (UID: "0fc90a98-5870-4ba2-8c31-0891019c9ac4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.914905 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.914955 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf27b\" (UniqueName: \"kubernetes.io/projected/0fc90a98-5870-4ba2-8c31-0891019c9ac4-kube-api-access-pf27b\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.914969 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc90a98-5870-4ba2-8c31-0891019c9ac4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.954715 4805 generic.go:334] "Generic (PLEG): container finished" podID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerID="cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a" exitCode=0 Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.954781 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fc90a98-5870-4ba2-8c31-0891019c9ac4","Type":"ContainerDied","Data":"cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a"} Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.954801 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.954825 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fc90a98-5870-4ba2-8c31-0891019c9ac4","Type":"ContainerDied","Data":"4a1e5de63d93c96b0da1c4a180d2725d843d4b971ac38fc773b5cdaca4a4dbb6"} Nov 28 15:47:56 crc kubenswrapper[4805]: I1128 15:47:56.954846 4805 scope.go:117] "RemoveContainer" containerID="cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.001788 4805 scope.go:117] "RemoveContainer" containerID="0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.009599 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.020291 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.029106 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:47:57 crc kubenswrapper[4805]: E1128 15:47:57.029484 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerName="nova-metadata-metadata" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.029503 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerName="nova-metadata-metadata" Nov 28 15:47:57 crc kubenswrapper[4805]: E1128 15:47:57.029530 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerName="nova-metadata-log" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.029537 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerName="nova-metadata-log" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.029701 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerName="nova-metadata-log" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.029725 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" containerName="nova-metadata-metadata" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.031942 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.035756 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.035791 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.042375 4805 scope.go:117] "RemoveContainer" containerID="cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.044422 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:47:57 crc kubenswrapper[4805]: E1128 15:47:57.050736 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a\": container with ID starting with cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a not found: ID does not exist" containerID="cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.050786 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a"} err="failed to get container status \"cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a\": rpc error: code = NotFound desc = could not find container \"cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a\": container with ID starting with cc470a460e07226ccb5b898f0e7ca95f56ba21ef5d0686cf6a68d81fdd73d46a not found: ID does not exist" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.050817 4805 scope.go:117] "RemoveContainer" containerID="0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f" Nov 28 15:47:57 crc kubenswrapper[4805]: E1128 15:47:57.053568 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f\": container with ID starting with 0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f not found: ID does not exist" containerID="0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.053620 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f"} err="failed to get container status \"0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f\": rpc error: code = NotFound desc = could not find container \"0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f\": container with ID starting with 0e98f5ba99014a68bb45794e508c955230c1dee719057475ec1a860864476c3f not found: ID does not exist" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.214114 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fc90a98-5870-4ba2-8c31-0891019c9ac4" path="/var/lib/kubelet/pods/0fc90a98-5870-4ba2-8c31-0891019c9ac4/volumes" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.222286 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-logs\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.222416 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-config-data\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.222491 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwfrp\" (UniqueName: \"kubernetes.io/projected/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-kube-api-access-rwfrp\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.222574 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.222640 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.323962 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-logs\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.324052 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-config-data\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.324108 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwfrp\" (UniqueName: \"kubernetes.io/projected/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-kube-api-access-rwfrp\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.324155 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.324199 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.324645 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-logs\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.339150 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-config-data\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.345347 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.347053 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.347717 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwfrp\" (UniqueName: \"kubernetes.io/projected/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-kube-api-access-rwfrp\") pod \"nova-metadata-0\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.352442 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.797338 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:47:57 crc kubenswrapper[4805]: I1128 15:47:57.965035 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f","Type":"ContainerStarted","Data":"09f3be79f117a469af928f8a404214270bf5d7978dea694f4746777d0176d7a7"} Nov 28 15:47:58 crc kubenswrapper[4805]: I1128 15:47:58.986918 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f","Type":"ContainerStarted","Data":"0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5"} Nov 28 15:47:58 crc kubenswrapper[4805]: I1128 15:47:58.987254 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f","Type":"ContainerStarted","Data":"3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a"} Nov 28 15:47:58 crc kubenswrapper[4805]: I1128 15:47:58.990023 4805 generic.go:334] "Generic (PLEG): container finished" podID="b6d86dfa-8c46-40bb-8155-438f6d5e06e6" containerID="c8dee08d7ace9297ae12c4f0d44cad8e080785d18e73355b29ebf5cc4b33729a" exitCode=0 Nov 28 15:47:58 crc kubenswrapper[4805]: I1128 15:47:58.990069 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bllqv" event={"ID":"b6d86dfa-8c46-40bb-8155-438f6d5e06e6","Type":"ContainerDied","Data":"c8dee08d7ace9297ae12c4f0d44cad8e080785d18e73355b29ebf5cc4b33729a"} Nov 28 15:47:59 crc kubenswrapper[4805]: I1128 15:47:59.016918 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.016897919 podStartE2EDuration="3.016897919s" podCreationTimestamp="2025-11-28 15:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:47:59.008446089 +0000 UTC m=+1306.058237450" watchObservedRunningTime="2025-11-28 15:47:59.016897919 +0000 UTC m=+1306.066689230" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.093023 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.130938 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.328283 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.427436 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.427499 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.476570 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.484798 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-config-data\") pod \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.484892 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-scripts\") pod \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.485046 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-combined-ca-bundle\") pod \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.485082 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxnj9\" (UniqueName: \"kubernetes.io/projected/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-kube-api-access-lxnj9\") pod \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\" (UID: \"b6d86dfa-8c46-40bb-8155-438f6d5e06e6\") " Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.492639 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-kube-api-access-lxnj9" (OuterVolumeSpecName: "kube-api-access-lxnj9") pod "b6d86dfa-8c46-40bb-8155-438f6d5e06e6" (UID: "b6d86dfa-8c46-40bb-8155-438f6d5e06e6"). InnerVolumeSpecName "kube-api-access-lxnj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.497516 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-scripts" (OuterVolumeSpecName: "scripts") pod "b6d86dfa-8c46-40bb-8155-438f6d5e06e6" (UID: "b6d86dfa-8c46-40bb-8155-438f6d5e06e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.535559 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-config-data" (OuterVolumeSpecName: "config-data") pod "b6d86dfa-8c46-40bb-8155-438f6d5e06e6" (UID: "b6d86dfa-8c46-40bb-8155-438f6d5e06e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.542437 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6d86dfa-8c46-40bb-8155-438f6d5e06e6" (UID: "b6d86dfa-8c46-40bb-8155-438f6d5e06e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.586575 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-mdsn8"] Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.587125 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" podUID="b77f479b-8072-4e56-8517-cce5ec74774c" containerName="dnsmasq-dns" containerID="cri-o://1d1558e326764d153e32793a2b7d65350d6db2ba68ebdad5bc0e57ce8100a927" gracePeriod=10 Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.589235 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.589267 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxnj9\" (UniqueName: \"kubernetes.io/projected/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-kube-api-access-lxnj9\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.589280 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:00 crc kubenswrapper[4805]: I1128 15:48:00.589291 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d86dfa-8c46-40bb-8155-438f6d5e06e6-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.013719 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bllqv" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.014447 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bllqv" event={"ID":"b6d86dfa-8c46-40bb-8155-438f6d5e06e6","Type":"ContainerDied","Data":"521f2d671001675d6624cd237bde3eb45a36d6137bcf47b7eba7184b9cc3b29c"} Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.014484 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="521f2d671001675d6624cd237bde3eb45a36d6137bcf47b7eba7184b9cc3b29c" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.017010 4805 generic.go:334] "Generic (PLEG): container finished" podID="b77f479b-8072-4e56-8517-cce5ec74774c" containerID="1d1558e326764d153e32793a2b7d65350d6db2ba68ebdad5bc0e57ce8100a927" exitCode=0 Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.018280 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" event={"ID":"b77f479b-8072-4e56-8517-cce5ec74774c","Type":"ContainerDied","Data":"1d1558e326764d153e32793a2b7d65350d6db2ba68ebdad5bc0e57ce8100a927"} Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.021469 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.077749 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.100976 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-sb\") pod \"b77f479b-8072-4e56-8517-cce5ec74774c\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.101045 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvjhx\" (UniqueName: \"kubernetes.io/projected/b77f479b-8072-4e56-8517-cce5ec74774c-kube-api-access-wvjhx\") pod \"b77f479b-8072-4e56-8517-cce5ec74774c\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.101250 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-config\") pod \"b77f479b-8072-4e56-8517-cce5ec74774c\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.101319 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-nb\") pod \"b77f479b-8072-4e56-8517-cce5ec74774c\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.101397 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-svc\") pod \"b77f479b-8072-4e56-8517-cce5ec74774c\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.101438 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-swift-storage-0\") pod \"b77f479b-8072-4e56-8517-cce5ec74774c\" (UID: \"b77f479b-8072-4e56-8517-cce5ec74774c\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.113606 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b77f479b-8072-4e56-8517-cce5ec74774c-kube-api-access-wvjhx" (OuterVolumeSpecName: "kube-api-access-wvjhx") pod "b77f479b-8072-4e56-8517-cce5ec74774c" (UID: "b77f479b-8072-4e56-8517-cce5ec74774c"). InnerVolumeSpecName "kube-api-access-wvjhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.172641 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b77f479b-8072-4e56-8517-cce5ec74774c" (UID: "b77f479b-8072-4e56-8517-cce5ec74774c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.173620 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b77f479b-8072-4e56-8517-cce5ec74774c" (UID: "b77f479b-8072-4e56-8517-cce5ec74774c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.175634 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-config" (OuterVolumeSpecName: "config") pod "b77f479b-8072-4e56-8517-cce5ec74774c" (UID: "b77f479b-8072-4e56-8517-cce5ec74774c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.176083 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b77f479b-8072-4e56-8517-cce5ec74774c" (UID: "b77f479b-8072-4e56-8517-cce5ec74774c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.181675 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b77f479b-8072-4e56-8517-cce5ec74774c" (UID: "b77f479b-8072-4e56-8517-cce5ec74774c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.204217 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.204274 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.204293 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.204310 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.204329 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b77f479b-8072-4e56-8517-cce5ec74774c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.204346 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvjhx\" (UniqueName: \"kubernetes.io/projected/b77f479b-8072-4e56-8517-cce5ec74774c-kube-api-access-wvjhx\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.222664 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.222903 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-log" containerID="cri-o://358443cbcb7f69cbca8e140908a93dcf6c47bd8d6c4023ba06d915c0a12242d9" gracePeriod=30 Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.222987 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-api" containerID="cri-o://ba50f2ffcc3947dbd305c2c3dbf7b3bea9c39237b598f372706ab0a19f01827c" gracePeriod=30 Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.227754 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": EOF" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.227754 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": EOF" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.241216 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.241440 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerName="nova-metadata-log" containerID="cri-o://3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a" gracePeriod=30 Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.241546 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerName="nova-metadata-metadata" containerID="cri-o://0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5" gracePeriod=30 Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.630380 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.727208 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.814201 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-nova-metadata-tls-certs\") pod \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.814327 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-config-data\") pod \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.814574 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwfrp\" (UniqueName: \"kubernetes.io/projected/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-kube-api-access-rwfrp\") pod \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.814751 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-combined-ca-bundle\") pod \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.814784 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-logs\") pod \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\" (UID: \"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f\") " Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.815873 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-logs" (OuterVolumeSpecName: "logs") pod "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" (UID: "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.820549 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-kube-api-access-rwfrp" (OuterVolumeSpecName: "kube-api-access-rwfrp") pod "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" (UID: "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f"). InnerVolumeSpecName "kube-api-access-rwfrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.847539 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" (UID: "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.850609 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-config-data" (OuterVolumeSpecName: "config-data") pod "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" (UID: "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.870941 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" (UID: "b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.917318 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.917377 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwfrp\" (UniqueName: \"kubernetes.io/projected/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-kube-api-access-rwfrp\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.917395 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.917406 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:01 crc kubenswrapper[4805]: I1128 15:48:01.917416 4805 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.025779 4805 generic.go:334] "Generic (PLEG): container finished" podID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerID="0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5" exitCode=0 Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.025812 4805 generic.go:334] "Generic (PLEG): container finished" podID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerID="3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a" exitCode=143 Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.025849 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f","Type":"ContainerDied","Data":"0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5"} Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.025875 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f","Type":"ContainerDied","Data":"3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a"} Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.025885 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f","Type":"ContainerDied","Data":"09f3be79f117a469af928f8a404214270bf5d7978dea694f4746777d0176d7a7"} Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.025899 4805 scope.go:117] "RemoveContainer" containerID="0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.026063 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.032230 4805 generic.go:334] "Generic (PLEG): container finished" podID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerID="358443cbcb7f69cbca8e140908a93dcf6c47bd8d6c4023ba06d915c0a12242d9" exitCode=143 Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.032320 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b7739baa-2c83-453a-9bb7-539cff74b9f7","Type":"ContainerDied","Data":"358443cbcb7f69cbca8e140908a93dcf6c47bd8d6c4023ba06d915c0a12242d9"} Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.035010 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.035420 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-mdsn8" event={"ID":"b77f479b-8072-4e56-8517-cce5ec74774c","Type":"ContainerDied","Data":"8822aa83abea6917bb9adf5f7a52fe9ccb62d1e60afabde9ce668bb84df84ecd"} Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.056627 4805 scope.go:117] "RemoveContainer" containerID="3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.069255 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.082084 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.093518 4805 scope.go:117] "RemoveContainer" containerID="0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5" Nov 28 15:48:02 crc kubenswrapper[4805]: E1128 15:48:02.094627 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5\": container with ID starting with 0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5 not found: ID does not exist" containerID="0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.094653 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5"} err="failed to get container status \"0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5\": rpc error: code = NotFound desc = could not find container \"0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5\": container with ID starting with 0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5 not found: ID does not exist" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.094675 4805 scope.go:117] "RemoveContainer" containerID="3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a" Nov 28 15:48:02 crc kubenswrapper[4805]: E1128 15:48:02.095267 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a\": container with ID starting with 3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a not found: ID does not exist" containerID="3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.095311 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a"} err="failed to get container status \"3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a\": rpc error: code = NotFound desc = could not find container \"3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a\": container with ID starting with 3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a not found: ID does not exist" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.095341 4805 scope.go:117] "RemoveContainer" containerID="0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.095654 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5"} err="failed to get container status \"0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5\": rpc error: code = NotFound desc = could not find container \"0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5\": container with ID starting with 0a7a7d55f2265da018b69467b2ae3a4bf2ceb2ae37c9d3ca7ecf18eb10e83bf5 not found: ID does not exist" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.095679 4805 scope.go:117] "RemoveContainer" containerID="3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.095906 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a"} err="failed to get container status \"3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a\": rpc error: code = NotFound desc = could not find container \"3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a\": container with ID starting with 3eb1552d03057e452bec1f1779732068d464633a20d7812e12b607306ac7db2a not found: ID does not exist" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.095927 4805 scope.go:117] "RemoveContainer" containerID="1d1558e326764d153e32793a2b7d65350d6db2ba68ebdad5bc0e57ce8100a927" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.102828 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-mdsn8"] Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.119788 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-mdsn8"] Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.124092 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:02 crc kubenswrapper[4805]: E1128 15:48:02.124567 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d86dfa-8c46-40bb-8155-438f6d5e06e6" containerName="nova-manage" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.124591 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d86dfa-8c46-40bb-8155-438f6d5e06e6" containerName="nova-manage" Nov 28 15:48:02 crc kubenswrapper[4805]: E1128 15:48:02.124607 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerName="nova-metadata-metadata" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.124615 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerName="nova-metadata-metadata" Nov 28 15:48:02 crc kubenswrapper[4805]: E1128 15:48:02.124651 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b77f479b-8072-4e56-8517-cce5ec74774c" containerName="init" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.124660 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b77f479b-8072-4e56-8517-cce5ec74774c" containerName="init" Nov 28 15:48:02 crc kubenswrapper[4805]: E1128 15:48:02.124688 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b77f479b-8072-4e56-8517-cce5ec74774c" containerName="dnsmasq-dns" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.124696 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b77f479b-8072-4e56-8517-cce5ec74774c" containerName="dnsmasq-dns" Nov 28 15:48:02 crc kubenswrapper[4805]: E1128 15:48:02.124709 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerName="nova-metadata-log" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.124716 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerName="nova-metadata-log" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.124987 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerName="nova-metadata-log" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.125017 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d86dfa-8c46-40bb-8155-438f6d5e06e6" containerName="nova-manage" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.125032 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" containerName="nova-metadata-metadata" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.125061 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b77f479b-8072-4e56-8517-cce5ec74774c" containerName="dnsmasq-dns" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.126317 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.127639 4805 scope.go:117] "RemoveContainer" containerID="bd5a19118f4db95a9c973e64adabb9e0dc85e4fbbc7a09353658f5767597a719" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.128554 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.128673 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.142585 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.224022 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-config-data\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.224102 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.224172 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed268974-2606-4a45-9e62-c2f54d728290-logs\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.224232 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.224248 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r75m9\" (UniqueName: \"kubernetes.io/projected/ed268974-2606-4a45-9e62-c2f54d728290-kube-api-access-r75m9\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.326269 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-config-data\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.326343 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.326443 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed268974-2606-4a45-9e62-c2f54d728290-logs\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.326541 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.326567 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r75m9\" (UniqueName: \"kubernetes.io/projected/ed268974-2606-4a45-9e62-c2f54d728290-kube-api-access-r75m9\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.328799 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed268974-2606-4a45-9e62-c2f54d728290-logs\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.332179 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.337169 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.337765 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-config-data\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.345810 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r75m9\" (UniqueName: \"kubernetes.io/projected/ed268974-2606-4a45-9e62-c2f54d728290-kube-api-access-r75m9\") pod \"nova-metadata-0\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " pod="openstack/nova-metadata-0" Nov 28 15:48:02 crc kubenswrapper[4805]: I1128 15:48:02.462127 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:48:03 crc kubenswrapper[4805]: I1128 15:48:03.039701 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:03 crc kubenswrapper[4805]: I1128 15:48:03.047340 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" containerName="nova-scheduler-scheduler" containerID="cri-o://0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8" gracePeriod=30 Nov 28 15:48:03 crc kubenswrapper[4805]: I1128 15:48:03.223485 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f" path="/var/lib/kubelet/pods/b4f5ddfb-1f5c-446e-b2e3-0a2465719e0f/volumes" Nov 28 15:48:03 crc kubenswrapper[4805]: I1128 15:48:03.225119 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b77f479b-8072-4e56-8517-cce5ec74774c" path="/var/lib/kubelet/pods/b77f479b-8072-4e56-8517-cce5ec74774c/volumes" Nov 28 15:48:04 crc kubenswrapper[4805]: I1128 15:48:04.056823 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed268974-2606-4a45-9e62-c2f54d728290","Type":"ContainerStarted","Data":"1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb"} Nov 28 15:48:04 crc kubenswrapper[4805]: I1128 15:48:04.057173 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed268974-2606-4a45-9e62-c2f54d728290","Type":"ContainerStarted","Data":"cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e"} Nov 28 15:48:04 crc kubenswrapper[4805]: I1128 15:48:04.057190 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed268974-2606-4a45-9e62-c2f54d728290","Type":"ContainerStarted","Data":"996248cc403d128f68af0bfb2ea6d46c5f5985677c60cab08f39616dbb4ee6b4"} Nov 28 15:48:04 crc kubenswrapper[4805]: I1128 15:48:04.058185 4805 generic.go:334] "Generic (PLEG): container finished" podID="39c6f2be-4e1d-40c1-8871-a008f39d2d49" containerID="93594d7686a712e66ead92d80c77d078b1f2f43f80e381c3d84546645e2f43fc" exitCode=0 Nov 28 15:48:04 crc kubenswrapper[4805]: I1128 15:48:04.058219 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pc5z4" event={"ID":"39c6f2be-4e1d-40c1-8871-a008f39d2d49","Type":"ContainerDied","Data":"93594d7686a712e66ead92d80c77d078b1f2f43f80e381c3d84546645e2f43fc"} Nov 28 15:48:04 crc kubenswrapper[4805]: I1128 15:48:04.090252 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.090231835 podStartE2EDuration="2.090231835s" podCreationTimestamp="2025-11-28 15:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:04.074667691 +0000 UTC m=+1311.124459042" watchObservedRunningTime="2025-11-28 15:48:04.090231835 +0000 UTC m=+1311.140023166" Nov 28 15:48:04 crc kubenswrapper[4805]: I1128 15:48:04.985191 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 15:48:05 crc kubenswrapper[4805]: E1128 15:48:05.092964 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8 is running failed: container process not found" containerID="0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 15:48:05 crc kubenswrapper[4805]: E1128 15:48:05.093341 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8 is running failed: container process not found" containerID="0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 15:48:05 crc kubenswrapper[4805]: E1128 15:48:05.093624 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8 is running failed: container process not found" containerID="0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 15:48:05 crc kubenswrapper[4805]: E1128 15:48:05.093661 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" containerName="nova-scheduler-scheduler" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.528366 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.535182 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.595427 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp4s2\" (UniqueName: \"kubernetes.io/projected/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-kube-api-access-jp4s2\") pod \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.595498 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khjd5\" (UniqueName: \"kubernetes.io/projected/39c6f2be-4e1d-40c1-8871-a008f39d2d49-kube-api-access-khjd5\") pod \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.595586 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-combined-ca-bundle\") pod \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.595669 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-config-data\") pod \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.595715 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-combined-ca-bundle\") pod \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.595738 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-scripts\") pod \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\" (UID: \"39c6f2be-4e1d-40c1-8871-a008f39d2d49\") " Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.595784 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-config-data\") pod \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\" (UID: \"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb\") " Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.602526 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-kube-api-access-jp4s2" (OuterVolumeSpecName: "kube-api-access-jp4s2") pod "6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" (UID: "6bbd17bb-4327-45e1-a9ad-6fd1294a62bb"). InnerVolumeSpecName "kube-api-access-jp4s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.603102 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39c6f2be-4e1d-40c1-8871-a008f39d2d49-kube-api-access-khjd5" (OuterVolumeSpecName: "kube-api-access-khjd5") pod "39c6f2be-4e1d-40c1-8871-a008f39d2d49" (UID: "39c6f2be-4e1d-40c1-8871-a008f39d2d49"). InnerVolumeSpecName "kube-api-access-khjd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.604640 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-scripts" (OuterVolumeSpecName: "scripts") pod "39c6f2be-4e1d-40c1-8871-a008f39d2d49" (UID: "39c6f2be-4e1d-40c1-8871-a008f39d2d49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.635955 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39c6f2be-4e1d-40c1-8871-a008f39d2d49" (UID: "39c6f2be-4e1d-40c1-8871-a008f39d2d49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.639240 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-config-data" (OuterVolumeSpecName: "config-data") pod "39c6f2be-4e1d-40c1-8871-a008f39d2d49" (UID: "39c6f2be-4e1d-40c1-8871-a008f39d2d49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.653328 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-config-data" (OuterVolumeSpecName: "config-data") pod "6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" (UID: "6bbd17bb-4327-45e1-a9ad-6fd1294a62bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.663258 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" (UID: "6bbd17bb-4327-45e1-a9ad-6fd1294a62bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.697592 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.697653 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.697666 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39c6f2be-4e1d-40c1-8871-a008f39d2d49-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.697674 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.697691 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp4s2\" (UniqueName: \"kubernetes.io/projected/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-kube-api-access-jp4s2\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.697701 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khjd5\" (UniqueName: \"kubernetes.io/projected/39c6f2be-4e1d-40c1-8871-a008f39d2d49-kube-api-access-khjd5\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:05 crc kubenswrapper[4805]: I1128 15:48:05.697709 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.090760 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pc5z4" event={"ID":"39c6f2be-4e1d-40c1-8871-a008f39d2d49","Type":"ContainerDied","Data":"03396b519d892da63f88c2cb02e47351ec96ea9afc54700a2a63f380098045fd"} Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.090800 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03396b519d892da63f88c2cb02e47351ec96ea9afc54700a2a63f380098045fd" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.090824 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pc5z4" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.092635 4805 generic.go:334] "Generic (PLEG): container finished" podID="6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" containerID="0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8" exitCode=0 Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.092664 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb","Type":"ContainerDied","Data":"0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8"} Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.092683 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6bbd17bb-4327-45e1-a9ad-6fd1294a62bb","Type":"ContainerDied","Data":"f516e2b11362eac8f0cf8b0be9c17f460c3bdf65050d542c4e738721a117392c"} Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.092700 4805 scope.go:117] "RemoveContainer" containerID="0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.092823 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.154184 4805 scope.go:117] "RemoveContainer" containerID="0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8" Nov 28 15:48:06 crc kubenswrapper[4805]: E1128 15:48:06.155642 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8\": container with ID starting with 0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8 not found: ID does not exist" containerID="0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.155681 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8"} err="failed to get container status \"0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8\": rpc error: code = NotFound desc = could not find container \"0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8\": container with ID starting with 0b0f6659b8f8006cdf6d595f8b432ad864ff2af969a0908329964d0238c145f8 not found: ID does not exist" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.165499 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.179113 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.192521 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:06 crc kubenswrapper[4805]: E1128 15:48:06.193041 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" containerName="nova-scheduler-scheduler" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.193058 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" containerName="nova-scheduler-scheduler" Nov 28 15:48:06 crc kubenswrapper[4805]: E1128 15:48:06.193092 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39c6f2be-4e1d-40c1-8871-a008f39d2d49" containerName="nova-cell1-conductor-db-sync" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.193103 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="39c6f2be-4e1d-40c1-8871-a008f39d2d49" containerName="nova-cell1-conductor-db-sync" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.193348 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="39c6f2be-4e1d-40c1-8871-a008f39d2d49" containerName="nova-cell1-conductor-db-sync" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.193432 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" containerName="nova-scheduler-scheduler" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.194279 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.199673 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.206898 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.222796 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.224156 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.227107 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.234039 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.312168 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-config-data\") pod \"nova-scheduler-0\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.312238 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvsg6\" (UniqueName: \"kubernetes.io/projected/73710567-5e80-472c-a38d-9c6281e5ad10-kube-api-access-fvsg6\") pod \"nova-scheduler-0\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.312341 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.312382 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xghst\" (UniqueName: \"kubernetes.io/projected/51493fc6-cc57-4bbf-a6a8-0864160ade3f-kube-api-access-xghst\") pod \"nova-cell1-conductor-0\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.312441 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.312470 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.414055 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.414123 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xghst\" (UniqueName: \"kubernetes.io/projected/51493fc6-cc57-4bbf-a6a8-0864160ade3f-kube-api-access-xghst\") pod \"nova-cell1-conductor-0\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.414215 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.414250 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.414287 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-config-data\") pod \"nova-scheduler-0\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.414330 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvsg6\" (UniqueName: \"kubernetes.io/projected/73710567-5e80-472c-a38d-9c6281e5ad10-kube-api-access-fvsg6\") pod \"nova-scheduler-0\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.418501 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.419390 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-config-data\") pod \"nova-scheduler-0\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.419742 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.428007 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.441017 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xghst\" (UniqueName: \"kubernetes.io/projected/51493fc6-cc57-4bbf-a6a8-0864160ade3f-kube-api-access-xghst\") pod \"nova-cell1-conductor-0\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.442563 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvsg6\" (UniqueName: \"kubernetes.io/projected/73710567-5e80-472c-a38d-9c6281e5ad10-kube-api-access-fvsg6\") pod \"nova-scheduler-0\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.519280 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.544716 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:06 crc kubenswrapper[4805]: I1128 15:48:06.998059 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:07 crc kubenswrapper[4805]: W1128 15:48:07.014319 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73710567_5e80_472c_a38d_9c6281e5ad10.slice/crio-75cc3c888ff72ed0b0fc6346b16493482d304b0805efbbd1c62859bcdce44214 WatchSource:0}: Error finding container 75cc3c888ff72ed0b0fc6346b16493482d304b0805efbbd1c62859bcdce44214: Status 404 returned error can't find the container with id 75cc3c888ff72ed0b0fc6346b16493482d304b0805efbbd1c62859bcdce44214 Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.100526 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 15:48:07 crc kubenswrapper[4805]: W1128 15:48:07.106353 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51493fc6_cc57_4bbf_a6a8_0864160ade3f.slice/crio-f02f5af79881f215334652ad50e645be8b7f4449bd67bc35806c5d03b49a41a0 WatchSource:0}: Error finding container f02f5af79881f215334652ad50e645be8b7f4449bd67bc35806c5d03b49a41a0: Status 404 returned error can't find the container with id f02f5af79881f215334652ad50e645be8b7f4449bd67bc35806c5d03b49a41a0 Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.107608 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73710567-5e80-472c-a38d-9c6281e5ad10","Type":"ContainerStarted","Data":"75cc3c888ff72ed0b0fc6346b16493482d304b0805efbbd1c62859bcdce44214"} Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.117766 4805 generic.go:334] "Generic (PLEG): container finished" podID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerID="ba50f2ffcc3947dbd305c2c3dbf7b3bea9c39237b598f372706ab0a19f01827c" exitCode=0 Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.117821 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b7739baa-2c83-453a-9bb7-539cff74b9f7","Type":"ContainerDied","Data":"ba50f2ffcc3947dbd305c2c3dbf7b3bea9c39237b598f372706ab0a19f01827c"} Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.218189 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bbd17bb-4327-45e1-a9ad-6fd1294a62bb" path="/var/lib/kubelet/pods/6bbd17bb-4327-45e1-a9ad-6fd1294a62bb/volumes" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.462534 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.462646 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.582001 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.737742 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-combined-ca-bundle\") pod \"b7739baa-2c83-453a-9bb7-539cff74b9f7\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.737787 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-config-data\") pod \"b7739baa-2c83-453a-9bb7-539cff74b9f7\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.737872 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7739baa-2c83-453a-9bb7-539cff74b9f7-logs\") pod \"b7739baa-2c83-453a-9bb7-539cff74b9f7\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.737927 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqfqn\" (UniqueName: \"kubernetes.io/projected/b7739baa-2c83-453a-9bb7-539cff74b9f7-kube-api-access-sqfqn\") pod \"b7739baa-2c83-453a-9bb7-539cff74b9f7\" (UID: \"b7739baa-2c83-453a-9bb7-539cff74b9f7\") " Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.738283 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7739baa-2c83-453a-9bb7-539cff74b9f7-logs" (OuterVolumeSpecName: "logs") pod "b7739baa-2c83-453a-9bb7-539cff74b9f7" (UID: "b7739baa-2c83-453a-9bb7-539cff74b9f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.738478 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7739baa-2c83-453a-9bb7-539cff74b9f7-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.750593 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7739baa-2c83-453a-9bb7-539cff74b9f7-kube-api-access-sqfqn" (OuterVolumeSpecName: "kube-api-access-sqfqn") pod "b7739baa-2c83-453a-9bb7-539cff74b9f7" (UID: "b7739baa-2c83-453a-9bb7-539cff74b9f7"). InnerVolumeSpecName "kube-api-access-sqfqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.774028 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7739baa-2c83-453a-9bb7-539cff74b9f7" (UID: "b7739baa-2c83-453a-9bb7-539cff74b9f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.778009 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-config-data" (OuterVolumeSpecName: "config-data") pod "b7739baa-2c83-453a-9bb7-539cff74b9f7" (UID: "b7739baa-2c83-453a-9bb7-539cff74b9f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.841088 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqfqn\" (UniqueName: \"kubernetes.io/projected/b7739baa-2c83-453a-9bb7-539cff74b9f7-kube-api-access-sqfqn\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.841168 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:07 crc kubenswrapper[4805]: I1128 15:48:07.841185 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7739baa-2c83-453a-9bb7-539cff74b9f7-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.130429 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b7739baa-2c83-453a-9bb7-539cff74b9f7","Type":"ContainerDied","Data":"3161af124e6754f6db4d0d8859e685ae207735e0861f1754fdfcfdb3ea662c99"} Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.130489 4805 scope.go:117] "RemoveContainer" containerID="ba50f2ffcc3947dbd305c2c3dbf7b3bea9c39237b598f372706ab0a19f01827c" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.130638 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.137105 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"51493fc6-cc57-4bbf-a6a8-0864160ade3f","Type":"ContainerStarted","Data":"273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1"} Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.137145 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"51493fc6-cc57-4bbf-a6a8-0864160ade3f","Type":"ContainerStarted","Data":"f02f5af79881f215334652ad50e645be8b7f4449bd67bc35806c5d03b49a41a0"} Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.137911 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.140446 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73710567-5e80-472c-a38d-9c6281e5ad10","Type":"ContainerStarted","Data":"3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6"} Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.154526 4805 scope.go:117] "RemoveContainer" containerID="358443cbcb7f69cbca8e140908a93dcf6c47bd8d6c4023ba06d915c0a12242d9" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.171067 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.171041951 podStartE2EDuration="2.171041951s" podCreationTimestamp="2025-11-28 15:48:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:08.161047239 +0000 UTC m=+1315.210838550" watchObservedRunningTime="2025-11-28 15:48:08.171041951 +0000 UTC m=+1315.220833262" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.186339 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.202986 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.224649 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:08 crc kubenswrapper[4805]: E1128 15:48:08.225145 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-log" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.225173 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-log" Nov 28 15:48:08 crc kubenswrapper[4805]: E1128 15:48:08.225204 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-api" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.225214 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-api" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.225464 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-log" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.225496 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" containerName="nova-api-api" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.226788 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.233121 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.235060 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.2350422930000002 podStartE2EDuration="2.235042293s" podCreationTimestamp="2025-11-28 15:48:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:08.196317329 +0000 UTC m=+1315.246108650" watchObservedRunningTime="2025-11-28 15:48:08.235042293 +0000 UTC m=+1315.284833604" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.258887 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.353283 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4zqb\" (UniqueName: \"kubernetes.io/projected/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-kube-api-access-p4zqb\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.354140 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-logs\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.354409 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.354544 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-config-data\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.456171 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.456250 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-config-data\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.456314 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4zqb\" (UniqueName: \"kubernetes.io/projected/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-kube-api-access-p4zqb\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.456395 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-logs\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.457051 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-logs\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.460302 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.460758 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-config-data\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.491711 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4zqb\" (UniqueName: \"kubernetes.io/projected/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-kube-api-access-p4zqb\") pod \"nova-api-0\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " pod="openstack/nova-api-0" Nov 28 15:48:08 crc kubenswrapper[4805]: I1128 15:48:08.555534 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:09 crc kubenswrapper[4805]: W1128 15:48:09.033586 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb80b5fa1_de6c_481d_b84a_5cadc6c5f48b.slice/crio-f161fe8ff97eb4b7f3c339317a9763062b2f3028a028e1bc66b66ab9bb7be5b9 WatchSource:0}: Error finding container f161fe8ff97eb4b7f3c339317a9763062b2f3028a028e1bc66b66ab9bb7be5b9: Status 404 returned error can't find the container with id f161fe8ff97eb4b7f3c339317a9763062b2f3028a028e1bc66b66ab9bb7be5b9 Nov 28 15:48:09 crc kubenswrapper[4805]: I1128 15:48:09.038793 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:09 crc kubenswrapper[4805]: I1128 15:48:09.162606 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:48:09 crc kubenswrapper[4805]: I1128 15:48:09.162845 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="fb35eb4d-a935-4c99-a3f9-e2fc22c32157" containerName="kube-state-metrics" containerID="cri-o://8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508" gracePeriod=30 Nov 28 15:48:09 crc kubenswrapper[4805]: I1128 15:48:09.171031 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b","Type":"ContainerStarted","Data":"f161fe8ff97eb4b7f3c339317a9763062b2f3028a028e1bc66b66ab9bb7be5b9"} Nov 28 15:48:09 crc kubenswrapper[4805]: I1128 15:48:09.218038 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7739baa-2c83-453a-9bb7-539cff74b9f7" path="/var/lib/kubelet/pods/b7739baa-2c83-453a-9bb7-539cff74b9f7/volumes" Nov 28 15:48:09 crc kubenswrapper[4805]: I1128 15:48:09.644533 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 15:48:09 crc kubenswrapper[4805]: I1128 15:48:09.781932 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkhd6\" (UniqueName: \"kubernetes.io/projected/fb35eb4d-a935-4c99-a3f9-e2fc22c32157-kube-api-access-tkhd6\") pod \"fb35eb4d-a935-4c99-a3f9-e2fc22c32157\" (UID: \"fb35eb4d-a935-4c99-a3f9-e2fc22c32157\") " Nov 28 15:48:09 crc kubenswrapper[4805]: I1128 15:48:09.787603 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb35eb4d-a935-4c99-a3f9-e2fc22c32157-kube-api-access-tkhd6" (OuterVolumeSpecName: "kube-api-access-tkhd6") pod "fb35eb4d-a935-4c99-a3f9-e2fc22c32157" (UID: "fb35eb4d-a935-4c99-a3f9-e2fc22c32157"). InnerVolumeSpecName "kube-api-access-tkhd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:09 crc kubenswrapper[4805]: I1128 15:48:09.884751 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkhd6\" (UniqueName: \"kubernetes.io/projected/fb35eb4d-a935-4c99-a3f9-e2fc22c32157-kube-api-access-tkhd6\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.187125 4805 generic.go:334] "Generic (PLEG): container finished" podID="fb35eb4d-a935-4c99-a3f9-e2fc22c32157" containerID="8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508" exitCode=2 Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.187213 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fb35eb4d-a935-4c99-a3f9-e2fc22c32157","Type":"ContainerDied","Data":"8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508"} Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.188879 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fb35eb4d-a935-4c99-a3f9-e2fc22c32157","Type":"ContainerDied","Data":"d9bc28c0f716c3c13619ee4f2c3427c35314207f620bdbb23394af50f502b1c3"} Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.187242 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.188944 4805 scope.go:117] "RemoveContainer" containerID="8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.192402 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b","Type":"ContainerStarted","Data":"e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60"} Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.192446 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b","Type":"ContainerStarted","Data":"b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5"} Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.213155 4805 scope.go:117] "RemoveContainer" containerID="8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508" Nov 28 15:48:10 crc kubenswrapper[4805]: E1128 15:48:10.213674 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508\": container with ID starting with 8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508 not found: ID does not exist" containerID="8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.213725 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508"} err="failed to get container status \"8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508\": rpc error: code = NotFound desc = could not find container \"8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508\": container with ID starting with 8ae8b396f87d05dce8083d8bc18eb0796c78e5b763149ba17d9d64fbd5751508 not found: ID does not exist" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.221079 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.221063526 podStartE2EDuration="2.221063526s" podCreationTimestamp="2025-11-28 15:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:10.214611141 +0000 UTC m=+1317.264402452" watchObservedRunningTime="2025-11-28 15:48:10.221063526 +0000 UTC m=+1317.270854837" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.239590 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.252255 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.260206 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:48:10 crc kubenswrapper[4805]: E1128 15:48:10.260724 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb35eb4d-a935-4c99-a3f9-e2fc22c32157" containerName="kube-state-metrics" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.260751 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb35eb4d-a935-4c99-a3f9-e2fc22c32157" containerName="kube-state-metrics" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.266149 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb35eb4d-a935-4c99-a3f9-e2fc22c32157" containerName="kube-state-metrics" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.270944 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.274758 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.297133 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.304472 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.401002 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.401310 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.401483 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.401653 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxdxp\" (UniqueName: \"kubernetes.io/projected/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-api-access-kxdxp\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.503207 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.503574 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.503725 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxdxp\" (UniqueName: \"kubernetes.io/projected/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-api-access-kxdxp\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.503813 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.507845 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.508045 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.518025 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.519009 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxdxp\" (UniqueName: \"kubernetes.io/projected/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-api-access-kxdxp\") pod \"kube-state-metrics-0\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " pod="openstack/kube-state-metrics-0" Nov 28 15:48:10 crc kubenswrapper[4805]: I1128 15:48:10.611463 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.059801 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.060183 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.060241 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.061131 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2d2222cce3bf61f8537971c6c7eb69d6ee1501220244e7d4ba5697a12f1d2d5"} pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.061206 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" containerID="cri-o://f2d2222cce3bf61f8537971c6c7eb69d6ee1501220244e7d4ba5697a12f1d2d5" gracePeriod=600 Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.087141 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:48:11 crc kubenswrapper[4805]: W1128 15:48:11.093515 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefb91a4_432a_4c5d_8d28_dee413c660c3.slice/crio-be518e9ace91a19bf38638deafca21d394a18966a9f0499d4f0f3adb3830e2ad WatchSource:0}: Error finding container be518e9ace91a19bf38638deafca21d394a18966a9f0499d4f0f3adb3830e2ad: Status 404 returned error can't find the container with id be518e9ace91a19bf38638deafca21d394a18966a9f0499d4f0f3adb3830e2ad Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.095632 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.162841 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.163212 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="ceilometer-central-agent" containerID="cri-o://76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f" gracePeriod=30 Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.163267 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="proxy-httpd" containerID="cri-o://6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e" gracePeriod=30 Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.163330 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="sg-core" containerID="cri-o://231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7" gracePeriod=30 Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.163400 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="ceilometer-notification-agent" containerID="cri-o://51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3" gracePeriod=30 Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.223969 4805 generic.go:334] "Generic (PLEG): container finished" podID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerID="f2d2222cce3bf61f8537971c6c7eb69d6ee1501220244e7d4ba5697a12f1d2d5" exitCode=0 Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.232213 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb35eb4d-a935-4c99-a3f9-e2fc22c32157" path="/var/lib/kubelet/pods/fb35eb4d-a935-4c99-a3f9-e2fc22c32157/volumes" Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.232876 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aefb91a4-432a-4c5d-8d28-dee413c660c3","Type":"ContainerStarted","Data":"be518e9ace91a19bf38638deafca21d394a18966a9f0499d4f0f3adb3830e2ad"} Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.232914 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerDied","Data":"f2d2222cce3bf61f8537971c6c7eb69d6ee1501220244e7d4ba5697a12f1d2d5"} Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.232953 4805 scope.go:117] "RemoveContainer" containerID="ada3fc8f488f7799cae44933973db15a3177119e0d7d2d4556df785008a6700d" Nov 28 15:48:11 crc kubenswrapper[4805]: I1128 15:48:11.520449 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.239168 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed"} Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.240593 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aefb91a4-432a-4c5d-8d28-dee413c660c3","Type":"ContainerStarted","Data":"af248f060f9652198f3fb50557ebfd252fb89de3f39e24268855915e1ebf72bd"} Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.240714 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.243926 4805 generic.go:334] "Generic (PLEG): container finished" podID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerID="6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e" exitCode=0 Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.243952 4805 generic.go:334] "Generic (PLEG): container finished" podID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerID="231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7" exitCode=2 Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.243963 4805 generic.go:334] "Generic (PLEG): container finished" podID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerID="76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f" exitCode=0 Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.243982 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerDied","Data":"6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e"} Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.244001 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerDied","Data":"231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7"} Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.244012 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerDied","Data":"76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f"} Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.277014 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.906726403 podStartE2EDuration="2.276992292s" podCreationTimestamp="2025-11-28 15:48:10 +0000 UTC" firstStartedPulling="2025-11-28 15:48:11.095384387 +0000 UTC m=+1318.145175698" lastFinishedPulling="2025-11-28 15:48:11.465650276 +0000 UTC m=+1318.515441587" observedRunningTime="2025-11-28 15:48:12.273795385 +0000 UTC m=+1319.323586716" watchObservedRunningTime="2025-11-28 15:48:12.276992292 +0000 UTC m=+1319.326783603" Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.463019 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 15:48:12 crc kubenswrapper[4805]: I1128 15:48:12.463367 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 15:48:13 crc kubenswrapper[4805]: I1128 15:48:13.477772 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.188:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 15:48:13 crc kubenswrapper[4805]: I1128 15:48:13.477802 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.188:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 15:48:14 crc kubenswrapper[4805]: E1128 15:48:14.515289 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7cb2c0c_ecfc_4eda_9d06_25e568535d43.slice/crio-conmon-51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3.scope\": RecentStats: unable to find data in memory cache]" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.181046 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.299611 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-run-httpd\") pod \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.299653 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgwdb\" (UniqueName: \"kubernetes.io/projected/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-kube-api-access-dgwdb\") pod \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.299724 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-scripts\") pod \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.299756 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-log-httpd\") pod \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.299841 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-sg-core-conf-yaml\") pod \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.299877 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-config-data\") pod \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.300008 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-combined-ca-bundle\") pod \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\" (UID: \"c7cb2c0c-ecfc-4eda-9d06-25e568535d43\") " Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.301424 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c7cb2c0c-ecfc-4eda-9d06-25e568535d43" (UID: "c7cb2c0c-ecfc-4eda-9d06-25e568535d43"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.305846 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c7cb2c0c-ecfc-4eda-9d06-25e568535d43" (UID: "c7cb2c0c-ecfc-4eda-9d06-25e568535d43"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.311629 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-kube-api-access-dgwdb" (OuterVolumeSpecName: "kube-api-access-dgwdb") pod "c7cb2c0c-ecfc-4eda-9d06-25e568535d43" (UID: "c7cb2c0c-ecfc-4eda-9d06-25e568535d43"). InnerVolumeSpecName "kube-api-access-dgwdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.316056 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-scripts" (OuterVolumeSpecName: "scripts") pod "c7cb2c0c-ecfc-4eda-9d06-25e568535d43" (UID: "c7cb2c0c-ecfc-4eda-9d06-25e568535d43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.329186 4805 generic.go:334] "Generic (PLEG): container finished" podID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerID="51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3" exitCode=0 Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.329229 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerDied","Data":"51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3"} Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.329264 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7cb2c0c-ecfc-4eda-9d06-25e568535d43","Type":"ContainerDied","Data":"1ed60d0035061e1aeb8e2238fdd14e3701a46141a1b732980041fa08471f8efb"} Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.329285 4805 scope.go:117] "RemoveContainer" containerID="6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.329283 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.340865 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c7cb2c0c-ecfc-4eda-9d06-25e568535d43" (UID: "c7cb2c0c-ecfc-4eda-9d06-25e568535d43"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.402795 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.402837 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgwdb\" (UniqueName: \"kubernetes.io/projected/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-kube-api-access-dgwdb\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.402855 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.402874 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.402892 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.410532 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7cb2c0c-ecfc-4eda-9d06-25e568535d43" (UID: "c7cb2c0c-ecfc-4eda-9d06-25e568535d43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.425348 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-config-data" (OuterVolumeSpecName: "config-data") pod "c7cb2c0c-ecfc-4eda-9d06-25e568535d43" (UID: "c7cb2c0c-ecfc-4eda-9d06-25e568535d43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.504315 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.504386 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cb2c0c-ecfc-4eda-9d06-25e568535d43-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.507567 4805 scope.go:117] "RemoveContainer" containerID="231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.527372 4805 scope.go:117] "RemoveContainer" containerID="51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.549509 4805 scope.go:117] "RemoveContainer" containerID="76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.577617 4805 scope.go:117] "RemoveContainer" containerID="6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e" Nov 28 15:48:15 crc kubenswrapper[4805]: E1128 15:48:15.578174 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e\": container with ID starting with 6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e not found: ID does not exist" containerID="6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.578217 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e"} err="failed to get container status \"6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e\": rpc error: code = NotFound desc = could not find container \"6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e\": container with ID starting with 6f34a737343fd35523004269e8e9d5ba06d8e04da64111bc610beea42660db0e not found: ID does not exist" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.578245 4805 scope.go:117] "RemoveContainer" containerID="231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7" Nov 28 15:48:15 crc kubenswrapper[4805]: E1128 15:48:15.578899 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7\": container with ID starting with 231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7 not found: ID does not exist" containerID="231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.578923 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7"} err="failed to get container status \"231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7\": rpc error: code = NotFound desc = could not find container \"231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7\": container with ID starting with 231f1c77053fc05dec633264f9713108e99d9c9cdf0cdd9a46d0fd2be1ae74f7 not found: ID does not exist" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.578936 4805 scope.go:117] "RemoveContainer" containerID="51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3" Nov 28 15:48:15 crc kubenswrapper[4805]: E1128 15:48:15.579213 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3\": container with ID starting with 51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3 not found: ID does not exist" containerID="51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.579231 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3"} err="failed to get container status \"51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3\": rpc error: code = NotFound desc = could not find container \"51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3\": container with ID starting with 51b4e897f3869ec795c42b902a5ec174d48f0c4af246e5a2d35875a2f5630ec3 not found: ID does not exist" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.579243 4805 scope.go:117] "RemoveContainer" containerID="76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f" Nov 28 15:48:15 crc kubenswrapper[4805]: E1128 15:48:15.579508 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f\": container with ID starting with 76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f not found: ID does not exist" containerID="76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.579535 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f"} err="failed to get container status \"76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f\": rpc error: code = NotFound desc = could not find container \"76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f\": container with ID starting with 76cf31d2d58ba1f5a510897d2189bb25bf0b88666965e3a310c809caa74d563f not found: ID does not exist" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.662986 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.672402 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.690781 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:15 crc kubenswrapper[4805]: E1128 15:48:15.691149 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="ceilometer-central-agent" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.691167 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="ceilometer-central-agent" Nov 28 15:48:15 crc kubenswrapper[4805]: E1128 15:48:15.691184 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="proxy-httpd" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.691191 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="proxy-httpd" Nov 28 15:48:15 crc kubenswrapper[4805]: E1128 15:48:15.691208 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="ceilometer-notification-agent" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.691214 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="ceilometer-notification-agent" Nov 28 15:48:15 crc kubenswrapper[4805]: E1128 15:48:15.691232 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="sg-core" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.691238 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="sg-core" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.691442 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="proxy-httpd" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.691465 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="ceilometer-central-agent" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.691486 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="ceilometer-notification-agent" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.691514 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" containerName="sg-core" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.697250 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.702091 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.702279 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.702447 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.746836 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.818822 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-config-data\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.818891 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nffh8\" (UniqueName: \"kubernetes.io/projected/bfb1257c-2a9e-4786-97f7-722493bb473a-kube-api-access-nffh8\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.818968 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.819020 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.819053 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-log-httpd\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.819118 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.823325 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-scripts\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.823520 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-run-httpd\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.924728 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-scripts\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.924794 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-run-httpd\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.924847 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-config-data\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.924895 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nffh8\" (UniqueName: \"kubernetes.io/projected/bfb1257c-2a9e-4786-97f7-722493bb473a-kube-api-access-nffh8\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.924942 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.924966 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.924985 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-log-httpd\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.925021 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.925603 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-run-httpd\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.925669 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-log-httpd\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.929778 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.930172 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.930325 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.930800 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-config-data\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.931497 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-scripts\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:15 crc kubenswrapper[4805]: I1128 15:48:15.941242 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nffh8\" (UniqueName: \"kubernetes.io/projected/bfb1257c-2a9e-4786-97f7-722493bb473a-kube-api-access-nffh8\") pod \"ceilometer-0\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " pod="openstack/ceilometer-0" Nov 28 15:48:16 crc kubenswrapper[4805]: I1128 15:48:16.024419 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:48:16 crc kubenswrapper[4805]: I1128 15:48:16.519651 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 15:48:16 crc kubenswrapper[4805]: I1128 15:48:16.544439 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:16 crc kubenswrapper[4805]: I1128 15:48:16.565649 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 15:48:16 crc kubenswrapper[4805]: W1128 15:48:16.566591 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfb1257c_2a9e_4786_97f7_722493bb473a.slice/crio-322b5d46edd550a2f3993a43aa8c565058abd4b9431d2a8f44e20e8153edfe2b WatchSource:0}: Error finding container 322b5d46edd550a2f3993a43aa8c565058abd4b9431d2a8f44e20e8153edfe2b: Status 404 returned error can't find the container with id 322b5d46edd550a2f3993a43aa8c565058abd4b9431d2a8f44e20e8153edfe2b Nov 28 15:48:16 crc kubenswrapper[4805]: I1128 15:48:16.591102 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 28 15:48:17 crc kubenswrapper[4805]: I1128 15:48:17.216647 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7cb2c0c-ecfc-4eda-9d06-25e568535d43" path="/var/lib/kubelet/pods/c7cb2c0c-ecfc-4eda-9d06-25e568535d43/volumes" Nov 28 15:48:17 crc kubenswrapper[4805]: I1128 15:48:17.368069 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerStarted","Data":"322b5d46edd550a2f3993a43aa8c565058abd4b9431d2a8f44e20e8153edfe2b"} Nov 28 15:48:17 crc kubenswrapper[4805]: I1128 15:48:17.396412 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 15:48:18 crc kubenswrapper[4805]: I1128 15:48:18.382643 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerStarted","Data":"ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0"} Nov 28 15:48:18 crc kubenswrapper[4805]: I1128 15:48:18.556182 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 15:48:18 crc kubenswrapper[4805]: I1128 15:48:18.556560 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 15:48:19 crc kubenswrapper[4805]: I1128 15:48:19.394633 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerStarted","Data":"0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea"} Nov 28 15:48:19 crc kubenswrapper[4805]: I1128 15:48:19.394956 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerStarted","Data":"7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1"} Nov 28 15:48:19 crc kubenswrapper[4805]: I1128 15:48:19.641695 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 15:48:19 crc kubenswrapper[4805]: I1128 15:48:19.641899 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 15:48:20 crc kubenswrapper[4805]: I1128 15:48:20.631781 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 28 15:48:21 crc kubenswrapper[4805]: I1128 15:48:21.415891 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerStarted","Data":"c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34"} Nov 28 15:48:21 crc kubenswrapper[4805]: I1128 15:48:21.416521 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 15:48:21 crc kubenswrapper[4805]: I1128 15:48:21.449149 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.312379533 podStartE2EDuration="6.449125623s" podCreationTimestamp="2025-11-28 15:48:15 +0000 UTC" firstStartedPulling="2025-11-28 15:48:16.569572523 +0000 UTC m=+1323.619363834" lastFinishedPulling="2025-11-28 15:48:20.706318613 +0000 UTC m=+1327.756109924" observedRunningTime="2025-11-28 15:48:21.436474899 +0000 UTC m=+1328.486266220" watchObservedRunningTime="2025-11-28 15:48:21.449125623 +0000 UTC m=+1328.498917014" Nov 28 15:48:22 crc kubenswrapper[4805]: I1128 15:48:22.470675 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 15:48:22 crc kubenswrapper[4805]: I1128 15:48:22.472674 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 15:48:22 crc kubenswrapper[4805]: I1128 15:48:22.480255 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 15:48:23 crc kubenswrapper[4805]: I1128 15:48:23.442598 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 15:48:25 crc kubenswrapper[4805]: I1128 15:48:25.458086 4805 generic.go:334] "Generic (PLEG): container finished" podID="7d2addf2-9fd9-4f75-804c-7668ca2684a3" containerID="dfeb9cf2753f320f8bad6ffc3d93be1753051da92f071ac9cfa6ec09ca20c114" exitCode=137 Nov 28 15:48:25 crc kubenswrapper[4805]: I1128 15:48:25.458146 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7d2addf2-9fd9-4f75-804c-7668ca2684a3","Type":"ContainerDied","Data":"dfeb9cf2753f320f8bad6ffc3d93be1753051da92f071ac9cfa6ec09ca20c114"} Nov 28 15:48:25 crc kubenswrapper[4805]: I1128 15:48:25.903097 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.051077 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-config-data\") pod \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.051269 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-combined-ca-bundle\") pod \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.052476 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vtrz\" (UniqueName: \"kubernetes.io/projected/7d2addf2-9fd9-4f75-804c-7668ca2684a3-kube-api-access-9vtrz\") pod \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\" (UID: \"7d2addf2-9fd9-4f75-804c-7668ca2684a3\") " Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.056782 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d2addf2-9fd9-4f75-804c-7668ca2684a3-kube-api-access-9vtrz" (OuterVolumeSpecName: "kube-api-access-9vtrz") pod "7d2addf2-9fd9-4f75-804c-7668ca2684a3" (UID: "7d2addf2-9fd9-4f75-804c-7668ca2684a3"). InnerVolumeSpecName "kube-api-access-9vtrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.091548 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d2addf2-9fd9-4f75-804c-7668ca2684a3" (UID: "7d2addf2-9fd9-4f75-804c-7668ca2684a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.091749 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-config-data" (OuterVolumeSpecName: "config-data") pod "7d2addf2-9fd9-4f75-804c-7668ca2684a3" (UID: "7d2addf2-9fd9-4f75-804c-7668ca2684a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.157067 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vtrz\" (UniqueName: \"kubernetes.io/projected/7d2addf2-9fd9-4f75-804c-7668ca2684a3-kube-api-access-9vtrz\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.157386 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.157409 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2addf2-9fd9-4f75-804c-7668ca2684a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.468878 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7d2addf2-9fd9-4f75-804c-7668ca2684a3","Type":"ContainerDied","Data":"7507b27a704e4c73b66fae6f62e63c2f8cc70f46c910935cb6bb3659dc88c867"} Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.468931 4805 scope.go:117] "RemoveContainer" containerID="dfeb9cf2753f320f8bad6ffc3d93be1753051da92f071ac9cfa6ec09ca20c114" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.468986 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.505409 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.513986 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.528176 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:48:26 crc kubenswrapper[4805]: E1128 15:48:26.528581 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d2addf2-9fd9-4f75-804c-7668ca2684a3" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.528600 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d2addf2-9fd9-4f75-804c-7668ca2684a3" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.528767 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d2addf2-9fd9-4f75-804c-7668ca2684a3" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.529371 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.535787 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.536216 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.540016 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.550189 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.566731 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.566886 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.567135 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.567179 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5tv6\" (UniqueName: \"kubernetes.io/projected/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-kube-api-access-h5tv6\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.567274 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.669469 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.669581 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.669674 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.669705 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5tv6\" (UniqueName: \"kubernetes.io/projected/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-kube-api-access-h5tv6\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.669749 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.674540 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.674958 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.675196 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.686492 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.690659 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5tv6\" (UniqueName: \"kubernetes.io/projected/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-kube-api-access-h5tv6\") pod \"nova-cell1-novncproxy-0\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:26 crc kubenswrapper[4805]: I1128 15:48:26.864517 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:27 crc kubenswrapper[4805]: I1128 15:48:27.224230 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d2addf2-9fd9-4f75-804c-7668ca2684a3" path="/var/lib/kubelet/pods/7d2addf2-9fd9-4f75-804c-7668ca2684a3/volumes" Nov 28 15:48:27 crc kubenswrapper[4805]: I1128 15:48:27.313683 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:48:27 crc kubenswrapper[4805]: W1128 15:48:27.314880 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode50653a2_4881_42fd_bc8a_35e5f8e2f48d.slice/crio-4a83f7ce8e72ffffef1ba1d028145e60ff20f88f8ded862d0fdde4857638014f WatchSource:0}: Error finding container 4a83f7ce8e72ffffef1ba1d028145e60ff20f88f8ded862d0fdde4857638014f: Status 404 returned error can't find the container with id 4a83f7ce8e72ffffef1ba1d028145e60ff20f88f8ded862d0fdde4857638014f Nov 28 15:48:27 crc kubenswrapper[4805]: I1128 15:48:27.480600 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e50653a2-4881-42fd-bc8a-35e5f8e2f48d","Type":"ContainerStarted","Data":"4a83f7ce8e72ffffef1ba1d028145e60ff20f88f8ded862d0fdde4857638014f"} Nov 28 15:48:28 crc kubenswrapper[4805]: I1128 15:48:28.498934 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e50653a2-4881-42fd-bc8a-35e5f8e2f48d","Type":"ContainerStarted","Data":"7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73"} Nov 28 15:48:28 crc kubenswrapper[4805]: I1128 15:48:28.538278 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.538253072 podStartE2EDuration="2.538253072s" podCreationTimestamp="2025-11-28 15:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:28.524986611 +0000 UTC m=+1335.574777942" watchObservedRunningTime="2025-11-28 15:48:28.538253072 +0000 UTC m=+1335.588044393" Nov 28 15:48:28 crc kubenswrapper[4805]: I1128 15:48:28.560647 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 15:48:28 crc kubenswrapper[4805]: I1128 15:48:28.561316 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 15:48:28 crc kubenswrapper[4805]: I1128 15:48:28.561984 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 15:48:28 crc kubenswrapper[4805]: I1128 15:48:28.565148 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.511063 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.514617 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.667290 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp"] Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.669586 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.683279 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp"] Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.832793 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-config\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.832855 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.832901 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-svc\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.832952 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqhb7\" (UniqueName: \"kubernetes.io/projected/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-kube-api-access-wqhb7\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.833002 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.833046 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.935486 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.935574 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-svc\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.935620 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqhb7\" (UniqueName: \"kubernetes.io/projected/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-kube-api-access-wqhb7\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.935707 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.935737 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.935812 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-config\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.936594 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-svc\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.936629 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.936685 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.936780 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-config\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.936867 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.960945 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqhb7\" (UniqueName: \"kubernetes.io/projected/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-kube-api-access-wqhb7\") pod \"dnsmasq-dns-5d8fc4ccc9-7jmkp\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:29 crc kubenswrapper[4805]: I1128 15:48:29.991947 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:30 crc kubenswrapper[4805]: I1128 15:48:30.510592 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp"] Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.536378 4805 generic.go:334] "Generic (PLEG): container finished" podID="b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" containerID="0c5162ba13d3afef2eddf48abd4934d9802217a483faa8ade07f7b373cd99503" exitCode=0 Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.538537 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" event={"ID":"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad","Type":"ContainerDied","Data":"0c5162ba13d3afef2eddf48abd4934d9802217a483faa8ade07f7b373cd99503"} Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.538605 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" event={"ID":"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad","Type":"ContainerStarted","Data":"af65fb6f9e2b6b005e2729ef460edbd9f67cab838751543c40ada167f3fdaa7d"} Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.665081 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.665416 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="ceilometer-central-agent" containerID="cri-o://ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0" gracePeriod=30 Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.666946 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="sg-core" containerID="cri-o://0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea" gracePeriod=30 Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.667103 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="proxy-httpd" containerID="cri-o://c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34" gracePeriod=30 Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.667164 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="ceilometer-notification-agent" containerID="cri-o://7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1" gracePeriod=30 Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.770133 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.193:3000/\": read tcp 10.217.0.2:45764->10.217.0.193:3000: read: connection reset by peer" Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.865263 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.933783 4805 scope.go:117] "RemoveContainer" containerID="32d027b8e872cc0576157c8ccf4cffa11b880c381b723c8b0c15b9505550afc4" Nov 28 15:48:31 crc kubenswrapper[4805]: I1128 15:48:31.957295 4805 scope.go:117] "RemoveContainer" containerID="5dcfa79e3056bcead728c317fef53686207b765501a5fb7e0d4f25d0b8edd449" Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.029742 4805 scope.go:117] "RemoveContainer" containerID="aa23e5b6e1d0bff09a5fab07983e8cdff900b8efff7108a269eb45026996547c" Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.311215 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.547207 4805 generic.go:334] "Generic (PLEG): container finished" podID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerID="c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34" exitCode=0 Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.547246 4805 generic.go:334] "Generic (PLEG): container finished" podID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerID="0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea" exitCode=2 Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.547257 4805 generic.go:334] "Generic (PLEG): container finished" podID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerID="ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0" exitCode=0 Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.547305 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerDied","Data":"c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34"} Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.547351 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerDied","Data":"0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea"} Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.547379 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerDied","Data":"ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0"} Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.549962 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" event={"ID":"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad","Type":"ContainerStarted","Data":"f512ffb73a4fae0efab8bd72a241efe3ae50f7565449d1f938e5a9c88c925c05"} Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.550008 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-log" containerID="cri-o://b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5" gracePeriod=30 Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.550115 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-api" containerID="cri-o://e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60" gracePeriod=30 Nov 28 15:48:32 crc kubenswrapper[4805]: I1128 15:48:32.589508 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" podStartSLOduration=3.589479644 podStartE2EDuration="3.589479644s" podCreationTimestamp="2025-11-28 15:48:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:32.574959259 +0000 UTC m=+1339.624750570" watchObservedRunningTime="2025-11-28 15:48:32.589479644 +0000 UTC m=+1339.639270955" Nov 28 15:48:33 crc kubenswrapper[4805]: I1128 15:48:33.564828 4805 generic.go:334] "Generic (PLEG): container finished" podID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerID="b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5" exitCode=143 Nov 28 15:48:33 crc kubenswrapper[4805]: I1128 15:48:33.564921 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b","Type":"ContainerDied","Data":"b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5"} Nov 28 15:48:33 crc kubenswrapper[4805]: I1128 15:48:33.565087 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.143557 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.343704 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-run-httpd\") pod \"bfb1257c-2a9e-4786-97f7-722493bb473a\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.344578 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-ceilometer-tls-certs\") pod \"bfb1257c-2a9e-4786-97f7-722493bb473a\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.344495 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bfb1257c-2a9e-4786-97f7-722493bb473a" (UID: "bfb1257c-2a9e-4786-97f7-722493bb473a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.344668 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-combined-ca-bundle\") pod \"bfb1257c-2a9e-4786-97f7-722493bb473a\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.345304 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-sg-core-conf-yaml\") pod \"bfb1257c-2a9e-4786-97f7-722493bb473a\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.345331 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nffh8\" (UniqueName: \"kubernetes.io/projected/bfb1257c-2a9e-4786-97f7-722493bb473a-kube-api-access-nffh8\") pod \"bfb1257c-2a9e-4786-97f7-722493bb473a\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.345386 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-scripts\") pod \"bfb1257c-2a9e-4786-97f7-722493bb473a\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.345407 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-config-data\") pod \"bfb1257c-2a9e-4786-97f7-722493bb473a\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.345933 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bfb1257c-2a9e-4786-97f7-722493bb473a" (UID: "bfb1257c-2a9e-4786-97f7-722493bb473a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.345561 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-log-httpd\") pod \"bfb1257c-2a9e-4786-97f7-722493bb473a\" (UID: \"bfb1257c-2a9e-4786-97f7-722493bb473a\") " Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.347089 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.347113 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfb1257c-2a9e-4786-97f7-722493bb473a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.351679 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb1257c-2a9e-4786-97f7-722493bb473a-kube-api-access-nffh8" (OuterVolumeSpecName: "kube-api-access-nffh8") pod "bfb1257c-2a9e-4786-97f7-722493bb473a" (UID: "bfb1257c-2a9e-4786-97f7-722493bb473a"). InnerVolumeSpecName "kube-api-access-nffh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.357150 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-scripts" (OuterVolumeSpecName: "scripts") pod "bfb1257c-2a9e-4786-97f7-722493bb473a" (UID: "bfb1257c-2a9e-4786-97f7-722493bb473a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.383048 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bfb1257c-2a9e-4786-97f7-722493bb473a" (UID: "bfb1257c-2a9e-4786-97f7-722493bb473a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.415600 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bfb1257c-2a9e-4786-97f7-722493bb473a" (UID: "bfb1257c-2a9e-4786-97f7-722493bb473a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.448789 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.448826 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nffh8\" (UniqueName: \"kubernetes.io/projected/bfb1257c-2a9e-4786-97f7-722493bb473a-kube-api-access-nffh8\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.448838 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.448848 4805 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.448935 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-config-data" (OuterVolumeSpecName: "config-data") pod "bfb1257c-2a9e-4786-97f7-722493bb473a" (UID: "bfb1257c-2a9e-4786-97f7-722493bb473a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.452313 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfb1257c-2a9e-4786-97f7-722493bb473a" (UID: "bfb1257c-2a9e-4786-97f7-722493bb473a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.550725 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.550768 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfb1257c-2a9e-4786-97f7-722493bb473a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.583736 4805 generic.go:334] "Generic (PLEG): container finished" podID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerID="7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1" exitCode=0 Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.583780 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerDied","Data":"7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1"} Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.583806 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfb1257c-2a9e-4786-97f7-722493bb473a","Type":"ContainerDied","Data":"322b5d46edd550a2f3993a43aa8c565058abd4b9431d2a8f44e20e8153edfe2b"} Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.583822 4805 scope.go:117] "RemoveContainer" containerID="c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.583929 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.610826 4805 scope.go:117] "RemoveContainer" containerID="0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.623932 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.635340 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.689691 4805 scope.go:117] "RemoveContainer" containerID="7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.692600 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:35 crc kubenswrapper[4805]: E1128 15:48:35.693266 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="ceilometer-central-agent" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.693284 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="ceilometer-central-agent" Nov 28 15:48:35 crc kubenswrapper[4805]: E1128 15:48:35.693314 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="sg-core" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.693322 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="sg-core" Nov 28 15:48:35 crc kubenswrapper[4805]: E1128 15:48:35.693351 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="ceilometer-notification-agent" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.693444 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="ceilometer-notification-agent" Nov 28 15:48:35 crc kubenswrapper[4805]: E1128 15:48:35.693472 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="proxy-httpd" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.693478 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="proxy-httpd" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.693794 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="ceilometer-central-agent" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.693819 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="sg-core" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.693843 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="ceilometer-notification-agent" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.693862 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" containerName="proxy-httpd" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.697747 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.708849 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.709092 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.709207 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.712537 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.717991 4805 scope.go:117] "RemoveContainer" containerID="ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.764342 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7clt7\" (UniqueName: \"kubernetes.io/projected/bd26b960-0994-4bb3-aef3-035519e32420-kube-api-access-7clt7\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.764519 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.764633 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.764746 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-run-httpd\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.764872 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-log-httpd\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.765192 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-config-data\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.765540 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.765791 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-scripts\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.867649 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.868045 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-scripts\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.868113 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7clt7\" (UniqueName: \"kubernetes.io/projected/bd26b960-0994-4bb3-aef3-035519e32420-kube-api-access-7clt7\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.868142 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.868173 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.868192 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-run-httpd\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.868213 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-log-httpd\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.868284 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-config-data\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.869059 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-run-httpd\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.869166 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-log-httpd\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.871958 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.872587 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-scripts\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.872937 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-config-data\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.880929 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.884511 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7clt7\" (UniqueName: \"kubernetes.io/projected/bd26b960-0994-4bb3-aef3-035519e32420-kube-api-access-7clt7\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.884967 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.942328 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.975405 4805 scope.go:117] "RemoveContainer" containerID="c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34" Nov 28 15:48:35 crc kubenswrapper[4805]: E1128 15:48:35.975919 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34\": container with ID starting with c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34 not found: ID does not exist" containerID="c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.976016 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34"} err="failed to get container status \"c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34\": rpc error: code = NotFound desc = could not find container \"c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34\": container with ID starting with c48fad1d5d14af4d9249f3e13aed43daabe2c01b8fb5165a551e1c7b9e211f34 not found: ID does not exist" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.976103 4805 scope.go:117] "RemoveContainer" containerID="0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea" Nov 28 15:48:35 crc kubenswrapper[4805]: E1128 15:48:35.976949 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea\": container with ID starting with 0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea not found: ID does not exist" containerID="0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.976981 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea"} err="failed to get container status \"0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea\": rpc error: code = NotFound desc = could not find container \"0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea\": container with ID starting with 0a41bcc2b979a3cbb4007ce3766506fd32612b3cac436d2ab3f3a14f34e383ea not found: ID does not exist" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.977007 4805 scope.go:117] "RemoveContainer" containerID="7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1" Nov 28 15:48:35 crc kubenswrapper[4805]: E1128 15:48:35.977259 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1\": container with ID starting with 7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1 not found: ID does not exist" containerID="7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.977326 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1"} err="failed to get container status \"7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1\": rpc error: code = NotFound desc = could not find container \"7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1\": container with ID starting with 7961d09f4099bb9fe0a67d59da8c2a0035e27486ce30eaaef7d8d122d97124b1 not found: ID does not exist" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.977347 4805 scope.go:117] "RemoveContainer" containerID="ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0" Nov 28 15:48:35 crc kubenswrapper[4805]: E1128 15:48:35.979038 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0\": container with ID starting with ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0 not found: ID does not exist" containerID="ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0" Nov 28 15:48:35 crc kubenswrapper[4805]: I1128 15:48:35.979083 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0"} err="failed to get container status \"ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0\": rpc error: code = NotFound desc = could not find container \"ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0\": container with ID starting with ebc796c600df14da9ccae16582b09e571de2e6108baa5356659d19e3f6a5a6e0 not found: ID does not exist" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.099004 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.174219 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-combined-ca-bundle\") pod \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.174401 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-logs\") pod \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.174440 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4zqb\" (UniqueName: \"kubernetes.io/projected/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-kube-api-access-p4zqb\") pod \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.174463 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-config-data\") pod \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\" (UID: \"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b\") " Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.175033 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-logs" (OuterVolumeSpecName: "logs") pod "b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" (UID: "b80b5fa1-de6c-481d-b84a-5cadc6c5f48b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.181591 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-kube-api-access-p4zqb" (OuterVolumeSpecName: "kube-api-access-p4zqb") pod "b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" (UID: "b80b5fa1-de6c-481d-b84a-5cadc6c5f48b"). InnerVolumeSpecName "kube-api-access-p4zqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.206469 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-config-data" (OuterVolumeSpecName: "config-data") pod "b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" (UID: "b80b5fa1-de6c-481d-b84a-5cadc6c5f48b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.215183 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" (UID: "b80b5fa1-de6c-481d-b84a-5cadc6c5f48b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.276806 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.276841 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.276854 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4zqb\" (UniqueName: \"kubernetes.io/projected/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-kube-api-access-p4zqb\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.276867 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.409976 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:48:36 crc kubenswrapper[4805]: W1128 15:48:36.412396 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd26b960_0994_4bb3_aef3_035519e32420.slice/crio-dcf4311d23e4526b4339c77ac45181acc6ef94781438d36bcef55e121805eb2c WatchSource:0}: Error finding container dcf4311d23e4526b4339c77ac45181acc6ef94781438d36bcef55e121805eb2c: Status 404 returned error can't find the container with id dcf4311d23e4526b4339c77ac45181acc6ef94781438d36bcef55e121805eb2c Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.593660 4805 generic.go:334] "Generic (PLEG): container finished" podID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerID="e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60" exitCode=0 Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.593707 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.593725 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b","Type":"ContainerDied","Data":"e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60"} Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.594120 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80b5fa1-de6c-481d-b84a-5cadc6c5f48b","Type":"ContainerDied","Data":"f161fe8ff97eb4b7f3c339317a9763062b2f3028a028e1bc66b66ab9bb7be5b9"} Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.594138 4805 scope.go:117] "RemoveContainer" containerID="e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.596964 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerStarted","Data":"dcf4311d23e4526b4339c77ac45181acc6ef94781438d36bcef55e121805eb2c"} Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.626156 4805 scope.go:117] "RemoveContainer" containerID="b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.629538 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.642024 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.650491 4805 scope.go:117] "RemoveContainer" containerID="e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60" Nov 28 15:48:36 crc kubenswrapper[4805]: E1128 15:48:36.650979 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60\": container with ID starting with e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60 not found: ID does not exist" containerID="e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.651027 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60"} err="failed to get container status \"e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60\": rpc error: code = NotFound desc = could not find container \"e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60\": container with ID starting with e530b55106fd5b92518f8c4e790fa28bd7b3fe065c808dcffed59a1d4b069f60 not found: ID does not exist" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.651062 4805 scope.go:117] "RemoveContainer" containerID="b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5" Nov 28 15:48:36 crc kubenswrapper[4805]: E1128 15:48:36.651402 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5\": container with ID starting with b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5 not found: ID does not exist" containerID="b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.651433 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5"} err="failed to get container status \"b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5\": rpc error: code = NotFound desc = could not find container \"b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5\": container with ID starting with b589ec122fda48eecd001fcaf782300371a08dfc4aab1e34d8d6524171dc39e5 not found: ID does not exist" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.653667 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:36 crc kubenswrapper[4805]: E1128 15:48:36.654092 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-api" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.654109 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-api" Nov 28 15:48:36 crc kubenswrapper[4805]: E1128 15:48:36.654128 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-log" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.654135 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-log" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.654299 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-log" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.654325 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" containerName="nova-api-api" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.655293 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.658519 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.658881 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.659057 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.664151 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.683960 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.684536 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-public-tls-certs\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.684681 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-config-data\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.684833 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5gl6\" (UniqueName: \"kubernetes.io/projected/e20a6a1d-d321-4646-945a-89d11180612d-kube-api-access-x5gl6\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.684991 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e20a6a1d-d321-4646-945a-89d11180612d-logs\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.685149 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.785891 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-public-tls-certs\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.785960 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-config-data\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.786023 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5gl6\" (UniqueName: \"kubernetes.io/projected/e20a6a1d-d321-4646-945a-89d11180612d-kube-api-access-x5gl6\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.786092 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e20a6a1d-d321-4646-945a-89d11180612d-logs\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.786121 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.786186 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.787501 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e20a6a1d-d321-4646-945a-89d11180612d-logs\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.790882 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.790961 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-config-data\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.791445 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.791685 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-public-tls-certs\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.808068 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5gl6\" (UniqueName: \"kubernetes.io/projected/e20a6a1d-d321-4646-945a-89d11180612d-kube-api-access-x5gl6\") pod \"nova-api-0\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " pod="openstack/nova-api-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.868172 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.892410 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:36 crc kubenswrapper[4805]: I1128 15:48:36.984217 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.227964 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b80b5fa1-de6c-481d-b84a-5cadc6c5f48b" path="/var/lib/kubelet/pods/b80b5fa1-de6c-481d-b84a-5cadc6c5f48b/volumes" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.230680 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfb1257c-2a9e-4786-97f7-722493bb473a" path="/var/lib/kubelet/pods/bfb1257c-2a9e-4786-97f7-722493bb473a/volumes" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.441676 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:37 crc kubenswrapper[4805]: W1128 15:48:37.447912 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode20a6a1d_d321_4646_945a_89d11180612d.slice/crio-81a9a308e749bef79fec379381058f87bef5ecfbc27d0e6ba66840e3bb67289a WatchSource:0}: Error finding container 81a9a308e749bef79fec379381058f87bef5ecfbc27d0e6ba66840e3bb67289a: Status 404 returned error can't find the container with id 81a9a308e749bef79fec379381058f87bef5ecfbc27d0e6ba66840e3bb67289a Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.614192 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerStarted","Data":"0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c"} Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.616660 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e20a6a1d-d321-4646-945a-89d11180612d","Type":"ContainerStarted","Data":"81a9a308e749bef79fec379381058f87bef5ecfbc27d0e6ba66840e3bb67289a"} Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.649995 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.839967 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-2sh5w"] Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.841509 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.843883 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.844094 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.848756 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2sh5w"] Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.919663 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.919895 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-config-data\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.919936 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg8xk\" (UniqueName: \"kubernetes.io/projected/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-kube-api-access-tg8xk\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:37 crc kubenswrapper[4805]: I1128 15:48:37.920048 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-scripts\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.021563 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-scripts\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.022454 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-config-data\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.022481 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.022542 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg8xk\" (UniqueName: \"kubernetes.io/projected/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-kube-api-access-tg8xk\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.027916 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-scripts\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.029637 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.038134 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-config-data\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.041971 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg8xk\" (UniqueName: \"kubernetes.io/projected/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-kube-api-access-tg8xk\") pod \"nova-cell1-cell-mapping-2sh5w\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.194078 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.635272 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e20a6a1d-d321-4646-945a-89d11180612d","Type":"ContainerStarted","Data":"35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903"} Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.635951 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e20a6a1d-d321-4646-945a-89d11180612d","Type":"ContainerStarted","Data":"054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7"} Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.639238 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerStarted","Data":"59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a"} Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.662417 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.66239335 podStartE2EDuration="2.66239335s" podCreationTimestamp="2025-11-28 15:48:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:38.653872968 +0000 UTC m=+1345.703664319" watchObservedRunningTime="2025-11-28 15:48:38.66239335 +0000 UTC m=+1345.712184671" Nov 28 15:48:38 crc kubenswrapper[4805]: W1128 15:48:38.702902 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda64cdf5b_5d0f_431f_a96f_ec0351559d2f.slice/crio-85eca1db2068acf1fc1cb37bcbfb479191f26733574a683c53e21224eccd0cfd WatchSource:0}: Error finding container 85eca1db2068acf1fc1cb37bcbfb479191f26733574a683c53e21224eccd0cfd: Status 404 returned error can't find the container with id 85eca1db2068acf1fc1cb37bcbfb479191f26733574a683c53e21224eccd0cfd Nov 28 15:48:38 crc kubenswrapper[4805]: I1128 15:48:38.703111 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2sh5w"] Nov 28 15:48:39 crc kubenswrapper[4805]: I1128 15:48:39.648757 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2sh5w" event={"ID":"a64cdf5b-5d0f-431f-a96f-ec0351559d2f","Type":"ContainerStarted","Data":"aae5a7230edb099b422ff386a01d34c04320a5e4d35f17b3aaf74b791db82ff0"} Nov 28 15:48:39 crc kubenswrapper[4805]: I1128 15:48:39.649118 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2sh5w" event={"ID":"a64cdf5b-5d0f-431f-a96f-ec0351559d2f","Type":"ContainerStarted","Data":"85eca1db2068acf1fc1cb37bcbfb479191f26733574a683c53e21224eccd0cfd"} Nov 28 15:48:39 crc kubenswrapper[4805]: I1128 15:48:39.652146 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerStarted","Data":"e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900"} Nov 28 15:48:39 crc kubenswrapper[4805]: I1128 15:48:39.667964 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-2sh5w" podStartSLOduration=2.667943722 podStartE2EDuration="2.667943722s" podCreationTimestamp="2025-11-28 15:48:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:39.661960579 +0000 UTC m=+1346.711751890" watchObservedRunningTime="2025-11-28 15:48:39.667943722 +0000 UTC m=+1346.717735033" Nov 28 15:48:39 crc kubenswrapper[4805]: I1128 15:48:39.993941 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.067825 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-8ln5n"] Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.068308 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" podUID="09e077c7-c008-4dae-af71-45fc22923773" containerName="dnsmasq-dns" containerID="cri-o://968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06" gracePeriod=10 Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.592618 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.667695 4805 generic.go:334] "Generic (PLEG): container finished" podID="09e077c7-c008-4dae-af71-45fc22923773" containerID="968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06" exitCode=0 Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.667807 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" event={"ID":"09e077c7-c008-4dae-af71-45fc22923773","Type":"ContainerDied","Data":"968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06"} Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.667843 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" event={"ID":"09e077c7-c008-4dae-af71-45fc22923773","Type":"ContainerDied","Data":"77e9d95935a04d8ed847507b1a68facebee44c7fdbfd574578a6027db33e65a5"} Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.667866 4805 scope.go:117] "RemoveContainer" containerID="968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.668051 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.693602 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerStarted","Data":"6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8"} Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.693659 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.696065 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-swift-storage-0\") pod \"09e077c7-c008-4dae-af71-45fc22923773\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.696112 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-config\") pod \"09e077c7-c008-4dae-af71-45fc22923773\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.696174 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-sb\") pod \"09e077c7-c008-4dae-af71-45fc22923773\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.696230 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6zjh\" (UniqueName: \"kubernetes.io/projected/09e077c7-c008-4dae-af71-45fc22923773-kube-api-access-g6zjh\") pod \"09e077c7-c008-4dae-af71-45fc22923773\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.696308 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-svc\") pod \"09e077c7-c008-4dae-af71-45fc22923773\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.696351 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-nb\") pod \"09e077c7-c008-4dae-af71-45fc22923773\" (UID: \"09e077c7-c008-4dae-af71-45fc22923773\") " Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.707288 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e077c7-c008-4dae-af71-45fc22923773-kube-api-access-g6zjh" (OuterVolumeSpecName: "kube-api-access-g6zjh") pod "09e077c7-c008-4dae-af71-45fc22923773" (UID: "09e077c7-c008-4dae-af71-45fc22923773"). InnerVolumeSpecName "kube-api-access-g6zjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.724934 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.10336942 podStartE2EDuration="5.724909114s" podCreationTimestamp="2025-11-28 15:48:35 +0000 UTC" firstStartedPulling="2025-11-28 15:48:36.41458445 +0000 UTC m=+1343.464375761" lastFinishedPulling="2025-11-28 15:48:40.036124144 +0000 UTC m=+1347.085915455" observedRunningTime="2025-11-28 15:48:40.712182618 +0000 UTC m=+1347.761973949" watchObservedRunningTime="2025-11-28 15:48:40.724909114 +0000 UTC m=+1347.774700425" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.735641 4805 scope.go:117] "RemoveContainer" containerID="a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.757773 4805 scope.go:117] "RemoveContainer" containerID="968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06" Nov 28 15:48:40 crc kubenswrapper[4805]: E1128 15:48:40.758766 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06\": container with ID starting with 968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06 not found: ID does not exist" containerID="968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.758826 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06"} err="failed to get container status \"968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06\": rpc error: code = NotFound desc = could not find container \"968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06\": container with ID starting with 968e9da605730c333a4198015c7d147814f618ca626141a1ad74bb86626b1b06 not found: ID does not exist" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.758862 4805 scope.go:117] "RemoveContainer" containerID="a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269" Nov 28 15:48:40 crc kubenswrapper[4805]: E1128 15:48:40.760043 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269\": container with ID starting with a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269 not found: ID does not exist" containerID="a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.760080 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269"} err="failed to get container status \"a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269\": rpc error: code = NotFound desc = could not find container \"a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269\": container with ID starting with a09a2357e453ffa2a10379a1b5827b60d7394787bb4205f1bc2667844a7a4269 not found: ID does not exist" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.762855 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-config" (OuterVolumeSpecName: "config") pod "09e077c7-c008-4dae-af71-45fc22923773" (UID: "09e077c7-c008-4dae-af71-45fc22923773"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.772646 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "09e077c7-c008-4dae-af71-45fc22923773" (UID: "09e077c7-c008-4dae-af71-45fc22923773"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.781776 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "09e077c7-c008-4dae-af71-45fc22923773" (UID: "09e077c7-c008-4dae-af71-45fc22923773"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.782034 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "09e077c7-c008-4dae-af71-45fc22923773" (UID: "09e077c7-c008-4dae-af71-45fc22923773"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.783452 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "09e077c7-c008-4dae-af71-45fc22923773" (UID: "09e077c7-c008-4dae-af71-45fc22923773"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.800174 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6zjh\" (UniqueName: \"kubernetes.io/projected/09e077c7-c008-4dae-af71-45fc22923773-kube-api-access-g6zjh\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.800517 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.800535 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.800548 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.800560 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:40 crc kubenswrapper[4805]: I1128 15:48:40.800571 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e077c7-c008-4dae-af71-45fc22923773-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:41 crc kubenswrapper[4805]: I1128 15:48:41.000775 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-8ln5n"] Nov 28 15:48:41 crc kubenswrapper[4805]: I1128 15:48:41.008925 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-8ln5n"] Nov 28 15:48:41 crc kubenswrapper[4805]: I1128 15:48:41.217406 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09e077c7-c008-4dae-af71-45fc22923773" path="/var/lib/kubelet/pods/09e077c7-c008-4dae-af71-45fc22923773/volumes" Nov 28 15:48:44 crc kubenswrapper[4805]: I1128 15:48:44.731292 4805 generic.go:334] "Generic (PLEG): container finished" podID="a64cdf5b-5d0f-431f-a96f-ec0351559d2f" containerID="aae5a7230edb099b422ff386a01d34c04320a5e4d35f17b3aaf74b791db82ff0" exitCode=0 Nov 28 15:48:44 crc kubenswrapper[4805]: I1128 15:48:44.731869 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2sh5w" event={"ID":"a64cdf5b-5d0f-431f-a96f-ec0351559d2f","Type":"ContainerDied","Data":"aae5a7230edb099b422ff386a01d34c04320a5e4d35f17b3aaf74b791db82ff0"} Nov 28 15:48:45 crc kubenswrapper[4805]: I1128 15:48:45.478255 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5594d9b959-8ln5n" podUID="09e077c7-c008-4dae-af71-45fc22923773" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.185:5353: i/o timeout" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.177732 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.299608 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-scripts\") pod \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.300087 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg8xk\" (UniqueName: \"kubernetes.io/projected/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-kube-api-access-tg8xk\") pod \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.300187 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-combined-ca-bundle\") pod \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.300215 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-config-data\") pod \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\" (UID: \"a64cdf5b-5d0f-431f-a96f-ec0351559d2f\") " Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.305926 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-scripts" (OuterVolumeSpecName: "scripts") pod "a64cdf5b-5d0f-431f-a96f-ec0351559d2f" (UID: "a64cdf5b-5d0f-431f-a96f-ec0351559d2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.319552 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-kube-api-access-tg8xk" (OuterVolumeSpecName: "kube-api-access-tg8xk") pod "a64cdf5b-5d0f-431f-a96f-ec0351559d2f" (UID: "a64cdf5b-5d0f-431f-a96f-ec0351559d2f"). InnerVolumeSpecName "kube-api-access-tg8xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.329563 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a64cdf5b-5d0f-431f-a96f-ec0351559d2f" (UID: "a64cdf5b-5d0f-431f-a96f-ec0351559d2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.342719 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-config-data" (OuterVolumeSpecName: "config-data") pod "a64cdf5b-5d0f-431f-a96f-ec0351559d2f" (UID: "a64cdf5b-5d0f-431f-a96f-ec0351559d2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.404022 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg8xk\" (UniqueName: \"kubernetes.io/projected/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-kube-api-access-tg8xk\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.404064 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.404073 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.404081 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64cdf5b-5d0f-431f-a96f-ec0351559d2f-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.750430 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2sh5w" event={"ID":"a64cdf5b-5d0f-431f-a96f-ec0351559d2f","Type":"ContainerDied","Data":"85eca1db2068acf1fc1cb37bcbfb479191f26733574a683c53e21224eccd0cfd"} Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.750489 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85eca1db2068acf1fc1cb37bcbfb479191f26733574a683c53e21224eccd0cfd" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.750501 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2sh5w" Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.953049 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.954212 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e20a6a1d-d321-4646-945a-89d11180612d" containerName="nova-api-api" containerID="cri-o://35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903" gracePeriod=30 Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.954230 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e20a6a1d-d321-4646-945a-89d11180612d" containerName="nova-api-log" containerID="cri-o://054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7" gracePeriod=30 Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.963848 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:46 crc kubenswrapper[4805]: I1128 15:48:46.964130 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="73710567-5e80-472c-a38d-9c6281e5ad10" containerName="nova-scheduler-scheduler" containerID="cri-o://3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6" gracePeriod=30 Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.048585 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.049108 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-log" containerID="cri-o://cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e" gracePeriod=30 Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.049318 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-metadata" containerID="cri-o://1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb" gracePeriod=30 Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.535935 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.725022 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5gl6\" (UniqueName: \"kubernetes.io/projected/e20a6a1d-d321-4646-945a-89d11180612d-kube-api-access-x5gl6\") pod \"e20a6a1d-d321-4646-945a-89d11180612d\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.725072 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-combined-ca-bundle\") pod \"e20a6a1d-d321-4646-945a-89d11180612d\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.725142 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-config-data\") pod \"e20a6a1d-d321-4646-945a-89d11180612d\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.725167 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e20a6a1d-d321-4646-945a-89d11180612d-logs\") pod \"e20a6a1d-d321-4646-945a-89d11180612d\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.725190 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-internal-tls-certs\") pod \"e20a6a1d-d321-4646-945a-89d11180612d\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.725273 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-public-tls-certs\") pod \"e20a6a1d-d321-4646-945a-89d11180612d\" (UID: \"e20a6a1d-d321-4646-945a-89d11180612d\") " Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.725993 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e20a6a1d-d321-4646-945a-89d11180612d-logs" (OuterVolumeSpecName: "logs") pod "e20a6a1d-d321-4646-945a-89d11180612d" (UID: "e20a6a1d-d321-4646-945a-89d11180612d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.731322 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e20a6a1d-d321-4646-945a-89d11180612d-kube-api-access-x5gl6" (OuterVolumeSpecName: "kube-api-access-x5gl6") pod "e20a6a1d-d321-4646-945a-89d11180612d" (UID: "e20a6a1d-d321-4646-945a-89d11180612d"). InnerVolumeSpecName "kube-api-access-x5gl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.758758 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-config-data" (OuterVolumeSpecName: "config-data") pod "e20a6a1d-d321-4646-945a-89d11180612d" (UID: "e20a6a1d-d321-4646-945a-89d11180612d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.762196 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e20a6a1d-d321-4646-945a-89d11180612d" (UID: "e20a6a1d-d321-4646-945a-89d11180612d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.762582 4805 generic.go:334] "Generic (PLEG): container finished" podID="ed268974-2606-4a45-9e62-c2f54d728290" containerID="cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e" exitCode=143 Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.762648 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed268974-2606-4a45-9e62-c2f54d728290","Type":"ContainerDied","Data":"cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e"} Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.764744 4805 generic.go:334] "Generic (PLEG): container finished" podID="e20a6a1d-d321-4646-945a-89d11180612d" containerID="35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903" exitCode=0 Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.764775 4805 generic.go:334] "Generic (PLEG): container finished" podID="e20a6a1d-d321-4646-945a-89d11180612d" containerID="054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7" exitCode=143 Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.764797 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e20a6a1d-d321-4646-945a-89d11180612d","Type":"ContainerDied","Data":"35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903"} Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.764820 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.764834 4805 scope.go:117] "RemoveContainer" containerID="35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.764821 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e20a6a1d-d321-4646-945a-89d11180612d","Type":"ContainerDied","Data":"054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7"} Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.764958 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e20a6a1d-d321-4646-945a-89d11180612d","Type":"ContainerDied","Data":"81a9a308e749bef79fec379381058f87bef5ecfbc27d0e6ba66840e3bb67289a"} Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.793521 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e20a6a1d-d321-4646-945a-89d11180612d" (UID: "e20a6a1d-d321-4646-945a-89d11180612d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.795046 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e20a6a1d-d321-4646-945a-89d11180612d" (UID: "e20a6a1d-d321-4646-945a-89d11180612d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.813046 4805 scope.go:117] "RemoveContainer" containerID="054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.828731 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5gl6\" (UniqueName: \"kubernetes.io/projected/e20a6a1d-d321-4646-945a-89d11180612d-kube-api-access-x5gl6\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.828771 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.828782 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.828793 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e20a6a1d-d321-4646-945a-89d11180612d-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.828806 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.828817 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20a6a1d-d321-4646-945a-89d11180612d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.836759 4805 scope.go:117] "RemoveContainer" containerID="35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903" Nov 28 15:48:47 crc kubenswrapper[4805]: E1128 15:48:47.837171 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903\": container with ID starting with 35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903 not found: ID does not exist" containerID="35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.837210 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903"} err="failed to get container status \"35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903\": rpc error: code = NotFound desc = could not find container \"35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903\": container with ID starting with 35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903 not found: ID does not exist" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.837236 4805 scope.go:117] "RemoveContainer" containerID="054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7" Nov 28 15:48:47 crc kubenswrapper[4805]: E1128 15:48:47.837531 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7\": container with ID starting with 054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7 not found: ID does not exist" containerID="054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.837564 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7"} err="failed to get container status \"054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7\": rpc error: code = NotFound desc = could not find container \"054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7\": container with ID starting with 054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7 not found: ID does not exist" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.837583 4805 scope.go:117] "RemoveContainer" containerID="35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.837833 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903"} err="failed to get container status \"35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903\": rpc error: code = NotFound desc = could not find container \"35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903\": container with ID starting with 35bb22089050c4015f34d3b0215843741c6d0a41ffb61b1c5fcf8fc83173e903 not found: ID does not exist" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.837873 4805 scope.go:117] "RemoveContainer" containerID="054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7" Nov 28 15:48:47 crc kubenswrapper[4805]: I1128 15:48:47.838090 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7"} err="failed to get container status \"054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7\": rpc error: code = NotFound desc = could not find container \"054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7\": container with ID starting with 054a17bf468f374dd38b3acb0eeab97b82ceec4747e7dc1043533bbc3f8eeba7 not found: ID does not exist" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.147070 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.157506 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.172691 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:48 crc kubenswrapper[4805]: E1128 15:48:48.173095 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20a6a1d-d321-4646-945a-89d11180612d" containerName="nova-api-api" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.173187 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20a6a1d-d321-4646-945a-89d11180612d" containerName="nova-api-api" Nov 28 15:48:48 crc kubenswrapper[4805]: E1128 15:48:48.173212 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e077c7-c008-4dae-af71-45fc22923773" containerName="dnsmasq-dns" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.173221 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e077c7-c008-4dae-af71-45fc22923773" containerName="dnsmasq-dns" Nov 28 15:48:48 crc kubenswrapper[4805]: E1128 15:48:48.173237 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20a6a1d-d321-4646-945a-89d11180612d" containerName="nova-api-log" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.173246 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20a6a1d-d321-4646-945a-89d11180612d" containerName="nova-api-log" Nov 28 15:48:48 crc kubenswrapper[4805]: E1128 15:48:48.173258 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64cdf5b-5d0f-431f-a96f-ec0351559d2f" containerName="nova-manage" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.173265 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64cdf5b-5d0f-431f-a96f-ec0351559d2f" containerName="nova-manage" Nov 28 15:48:48 crc kubenswrapper[4805]: E1128 15:48:48.173280 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e077c7-c008-4dae-af71-45fc22923773" containerName="init" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.173287 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e077c7-c008-4dae-af71-45fc22923773" containerName="init" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.173570 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64cdf5b-5d0f-431f-a96f-ec0351559d2f" containerName="nova-manage" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.173595 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e20a6a1d-d321-4646-945a-89d11180612d" containerName="nova-api-log" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.173621 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e20a6a1d-d321-4646-945a-89d11180612d" containerName="nova-api-api" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.173637 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="09e077c7-c008-4dae-af71-45fc22923773" containerName="dnsmasq-dns" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.174930 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.177170 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.178175 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.178194 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.198997 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.337911 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrqwf\" (UniqueName: \"kubernetes.io/projected/6ef711de-53e6-4705-8d2f-469b6dc2d4de-kube-api-access-qrqwf\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.337960 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ef711de-53e6-4705-8d2f-469b6dc2d4de-logs\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.337984 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.338067 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.338392 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-public-tls-certs\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.338508 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-config-data\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.440472 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrqwf\" (UniqueName: \"kubernetes.io/projected/6ef711de-53e6-4705-8d2f-469b6dc2d4de-kube-api-access-qrqwf\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.440840 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ef711de-53e6-4705-8d2f-469b6dc2d4de-logs\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.440882 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.440962 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.441035 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-public-tls-certs\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.441065 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-config-data\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.441234 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ef711de-53e6-4705-8d2f-469b6dc2d4de-logs\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.446119 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.446185 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-config-data\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.447478 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.448407 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-public-tls-certs\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.462889 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrqwf\" (UniqueName: \"kubernetes.io/projected/6ef711de-53e6-4705-8d2f-469b6dc2d4de-kube-api-access-qrqwf\") pod \"nova-api-0\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " pod="openstack/nova-api-0" Nov 28 15:48:48 crc kubenswrapper[4805]: I1128 15:48:48.499221 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:48:49 crc kubenswrapper[4805]: I1128 15:48:48.999704 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:48:49 crc kubenswrapper[4805]: W1128 15:48:49.003515 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ef711de_53e6_4705_8d2f_469b6dc2d4de.slice/crio-594f4fffd996d60bb888468f4ca60b1e16c97ad5601423a91ec7c755ceda4f9f WatchSource:0}: Error finding container 594f4fffd996d60bb888468f4ca60b1e16c97ad5601423a91ec7c755ceda4f9f: Status 404 returned error can't find the container with id 594f4fffd996d60bb888468f4ca60b1e16c97ad5601423a91ec7c755ceda4f9f Nov 28 15:48:49 crc kubenswrapper[4805]: I1128 15:48:49.218293 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e20a6a1d-d321-4646-945a-89d11180612d" path="/var/lib/kubelet/pods/e20a6a1d-d321-4646-945a-89d11180612d/volumes" Nov 28 15:48:49 crc kubenswrapper[4805]: I1128 15:48:49.788475 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ef711de-53e6-4705-8d2f-469b6dc2d4de","Type":"ContainerStarted","Data":"dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649"} Nov 28 15:48:49 crc kubenswrapper[4805]: I1128 15:48:49.788784 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ef711de-53e6-4705-8d2f-469b6dc2d4de","Type":"ContainerStarted","Data":"58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194"} Nov 28 15:48:49 crc kubenswrapper[4805]: I1128 15:48:49.788794 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ef711de-53e6-4705-8d2f-469b6dc2d4de","Type":"ContainerStarted","Data":"594f4fffd996d60bb888468f4ca60b1e16c97ad5601423a91ec7c755ceda4f9f"} Nov 28 15:48:49 crc kubenswrapper[4805]: I1128 15:48:49.826505 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.8264837250000001 podStartE2EDuration="1.826483725s" podCreationTimestamp="2025-11-28 15:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:49.814752556 +0000 UTC m=+1356.864543867" watchObservedRunningTime="2025-11-28 15:48:49.826483725 +0000 UTC m=+1356.876275046" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.206827 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.188:8775/\": read tcp 10.217.0.2:51108->10.217.0.188:8775: read: connection reset by peer" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.206859 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.188:8775/\": read tcp 10.217.0.2:51112->10.217.0.188:8775: read: connection reset by peer" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.717938 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.799222 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.799249 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed268974-2606-4a45-9e62-c2f54d728290","Type":"ContainerDied","Data":"1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb"} Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.799306 4805 scope.go:117] "RemoveContainer" containerID="1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.799315 4805 generic.go:334] "Generic (PLEG): container finished" podID="ed268974-2606-4a45-9e62-c2f54d728290" containerID="1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb" exitCode=0 Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.799445 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed268974-2606-4a45-9e62-c2f54d728290","Type":"ContainerDied","Data":"996248cc403d128f68af0bfb2ea6d46c5f5985677c60cab08f39616dbb4ee6b4"} Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.835329 4805 scope.go:117] "RemoveContainer" containerID="cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.884624 4805 scope.go:117] "RemoveContainer" containerID="1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb" Nov 28 15:48:50 crc kubenswrapper[4805]: E1128 15:48:50.885389 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb\": container with ID starting with 1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb not found: ID does not exist" containerID="1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.885430 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb"} err="failed to get container status \"1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb\": rpc error: code = NotFound desc = could not find container \"1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb\": container with ID starting with 1878b6c15e5ef9c2ae65388b98bc4f733dfd785fdeae4b2d17365e14ec76eceb not found: ID does not exist" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.885453 4805 scope.go:117] "RemoveContainer" containerID="cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e" Nov 28 15:48:50 crc kubenswrapper[4805]: E1128 15:48:50.885815 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e\": container with ID starting with cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e not found: ID does not exist" containerID="cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.885841 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e"} err="failed to get container status \"cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e\": rpc error: code = NotFound desc = could not find container \"cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e\": container with ID starting with cebc0ee774581e9c1b79eb46da0ae7d6818d407baab35367163857c0a3ace60e not found: ID does not exist" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.888710 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-config-data\") pod \"ed268974-2606-4a45-9e62-c2f54d728290\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.888758 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-combined-ca-bundle\") pod \"ed268974-2606-4a45-9e62-c2f54d728290\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.888822 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed268974-2606-4a45-9e62-c2f54d728290-logs\") pod \"ed268974-2606-4a45-9e62-c2f54d728290\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.888882 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-nova-metadata-tls-certs\") pod \"ed268974-2606-4a45-9e62-c2f54d728290\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.888923 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r75m9\" (UniqueName: \"kubernetes.io/projected/ed268974-2606-4a45-9e62-c2f54d728290-kube-api-access-r75m9\") pod \"ed268974-2606-4a45-9e62-c2f54d728290\" (UID: \"ed268974-2606-4a45-9e62-c2f54d728290\") " Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.890175 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed268974-2606-4a45-9e62-c2f54d728290-logs" (OuterVolumeSpecName: "logs") pod "ed268974-2606-4a45-9e62-c2f54d728290" (UID: "ed268974-2606-4a45-9e62-c2f54d728290"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.895118 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed268974-2606-4a45-9e62-c2f54d728290-kube-api-access-r75m9" (OuterVolumeSpecName: "kube-api-access-r75m9") pod "ed268974-2606-4a45-9e62-c2f54d728290" (UID: "ed268974-2606-4a45-9e62-c2f54d728290"). InnerVolumeSpecName "kube-api-access-r75m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.918112 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed268974-2606-4a45-9e62-c2f54d728290" (UID: "ed268974-2606-4a45-9e62-c2f54d728290"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.923170 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-config-data" (OuterVolumeSpecName: "config-data") pod "ed268974-2606-4a45-9e62-c2f54d728290" (UID: "ed268974-2606-4a45-9e62-c2f54d728290"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.960874 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ed268974-2606-4a45-9e62-c2f54d728290" (UID: "ed268974-2606-4a45-9e62-c2f54d728290"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.992725 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.992765 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.992780 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed268974-2606-4a45-9e62-c2f54d728290-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.992792 4805 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed268974-2606-4a45-9e62-c2f54d728290-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:50 crc kubenswrapper[4805]: I1128 15:48:50.992804 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r75m9\" (UniqueName: \"kubernetes.io/projected/ed268974-2606-4a45-9e62-c2f54d728290-kube-api-access-r75m9\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.144237 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.165266 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.174667 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:51 crc kubenswrapper[4805]: E1128 15:48:51.175519 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-log" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.175553 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-log" Nov 28 15:48:51 crc kubenswrapper[4805]: E1128 15:48:51.175574 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-metadata" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.175584 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-metadata" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.175920 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-metadata" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.175968 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed268974-2606-4a45-9e62-c2f54d728290" containerName="nova-metadata-log" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.177612 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.181955 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.182146 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.219187 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed268974-2606-4a45-9e62-c2f54d728290" path="/var/lib/kubelet/pods/ed268974-2606-4a45-9e62-c2f54d728290/volumes" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.220070 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.222286 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.301477 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-config-data\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.301543 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.301691 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7348256-32de-4696-9637-96d7323831ed-logs\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.301748 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.301887 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x45zf\" (UniqueName: \"kubernetes.io/projected/d7348256-32de-4696-9637-96d7323831ed-kube-api-access-x45zf\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.403202 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-combined-ca-bundle\") pod \"73710567-5e80-472c-a38d-9c6281e5ad10\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.403256 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-config-data\") pod \"73710567-5e80-472c-a38d-9c6281e5ad10\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.403387 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvsg6\" (UniqueName: \"kubernetes.io/projected/73710567-5e80-472c-a38d-9c6281e5ad10-kube-api-access-fvsg6\") pod \"73710567-5e80-472c-a38d-9c6281e5ad10\" (UID: \"73710567-5e80-472c-a38d-9c6281e5ad10\") " Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.403694 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.403830 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x45zf\" (UniqueName: \"kubernetes.io/projected/d7348256-32de-4696-9637-96d7323831ed-kube-api-access-x45zf\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.403859 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-config-data\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.403892 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.403952 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7348256-32de-4696-9637-96d7323831ed-logs\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.404430 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7348256-32de-4696-9637-96d7323831ed-logs\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.410059 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73710567-5e80-472c-a38d-9c6281e5ad10-kube-api-access-fvsg6" (OuterVolumeSpecName: "kube-api-access-fvsg6") pod "73710567-5e80-472c-a38d-9c6281e5ad10" (UID: "73710567-5e80-472c-a38d-9c6281e5ad10"). InnerVolumeSpecName "kube-api-access-fvsg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.410474 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.410723 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.413528 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-config-data\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.423876 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x45zf\" (UniqueName: \"kubernetes.io/projected/d7348256-32de-4696-9637-96d7323831ed-kube-api-access-x45zf\") pod \"nova-metadata-0\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.440883 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73710567-5e80-472c-a38d-9c6281e5ad10" (UID: "73710567-5e80-472c-a38d-9c6281e5ad10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.445869 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-config-data" (OuterVolumeSpecName: "config-data") pod "73710567-5e80-472c-a38d-9c6281e5ad10" (UID: "73710567-5e80-472c-a38d-9c6281e5ad10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.503999 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.505370 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvsg6\" (UniqueName: \"kubernetes.io/projected/73710567-5e80-472c-a38d-9c6281e5ad10-kube-api-access-fvsg6\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.505397 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.505406 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73710567-5e80-472c-a38d-9c6281e5ad10-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.813139 4805 generic.go:334] "Generic (PLEG): container finished" podID="73710567-5e80-472c-a38d-9c6281e5ad10" containerID="3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6" exitCode=0 Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.813222 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.813244 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73710567-5e80-472c-a38d-9c6281e5ad10","Type":"ContainerDied","Data":"3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6"} Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.814394 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73710567-5e80-472c-a38d-9c6281e5ad10","Type":"ContainerDied","Data":"75cc3c888ff72ed0b0fc6346b16493482d304b0805efbbd1c62859bcdce44214"} Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.814418 4805 scope.go:117] "RemoveContainer" containerID="3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.859314 4805 scope.go:117] "RemoveContainer" containerID="3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6" Nov 28 15:48:51 crc kubenswrapper[4805]: E1128 15:48:51.867937 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6\": container with ID starting with 3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6 not found: ID does not exist" containerID="3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.867995 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6"} err="failed to get container status \"3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6\": rpc error: code = NotFound desc = could not find container \"3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6\": container with ID starting with 3e60221881ab7e6abf2d34724322282328014954bb0d1173dc79dc5f17285ab6 not found: ID does not exist" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.882046 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.897839 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.914568 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:51 crc kubenswrapper[4805]: E1128 15:48:51.914991 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73710567-5e80-472c-a38d-9c6281e5ad10" containerName="nova-scheduler-scheduler" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.915011 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="73710567-5e80-472c-a38d-9c6281e5ad10" containerName="nova-scheduler-scheduler" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.915223 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="73710567-5e80-472c-a38d-9c6281e5ad10" containerName="nova-scheduler-scheduler" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.915917 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.921810 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.927504 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:51 crc kubenswrapper[4805]: I1128 15:48:51.964294 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:48:51 crc kubenswrapper[4805]: W1128 15:48:51.965894 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7348256_32de_4696_9637_96d7323831ed.slice/crio-75571db30208cbfdded485fc7f0b84911c3aaf764990bcced6068453ceb1e3da WatchSource:0}: Error finding container 75571db30208cbfdded485fc7f0b84911c3aaf764990bcced6068453ceb1e3da: Status 404 returned error can't find the container with id 75571db30208cbfdded485fc7f0b84911c3aaf764990bcced6068453ceb1e3da Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.013520 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.013596 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-config-data\") pod \"nova-scheduler-0\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.013652 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl9np\" (UniqueName: \"kubernetes.io/projected/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-kube-api-access-dl9np\") pod \"nova-scheduler-0\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.115258 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.115685 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-config-data\") pod \"nova-scheduler-0\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.115726 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl9np\" (UniqueName: \"kubernetes.io/projected/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-kube-api-access-dl9np\") pod \"nova-scheduler-0\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.121007 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-config-data\") pod \"nova-scheduler-0\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.122152 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.133022 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl9np\" (UniqueName: \"kubernetes.io/projected/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-kube-api-access-dl9np\") pod \"nova-scheduler-0\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.241443 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.704844 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.829895 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7348256-32de-4696-9637-96d7323831ed","Type":"ContainerStarted","Data":"5b7e15d112969a7b953c22fe09274e56dd365f74ce49e846d9443a14feb3ce7a"} Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.829941 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7348256-32de-4696-9637-96d7323831ed","Type":"ContainerStarted","Data":"099d02bc7f77ee26f120af8f3d1407c72c14067b902a1e28ada5fa2095d8e0a1"} Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.829950 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7348256-32de-4696-9637-96d7323831ed","Type":"ContainerStarted","Data":"75571db30208cbfdded485fc7f0b84911c3aaf764990bcced6068453ceb1e3da"} Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.832135 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ac086fdf-9a1a-40ab-ab2f-8a33dae28290","Type":"ContainerStarted","Data":"b5e2786e52c751fdc5d474871b9e78d5a30793ccb2a30907a1d2491c251f095d"} Nov 28 15:48:52 crc kubenswrapper[4805]: I1128 15:48:52.850776 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.85075894 podStartE2EDuration="1.85075894s" podCreationTimestamp="2025-11-28 15:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:52.846375042 +0000 UTC m=+1359.896166353" watchObservedRunningTime="2025-11-28 15:48:52.85075894 +0000 UTC m=+1359.900550251" Nov 28 15:48:53 crc kubenswrapper[4805]: I1128 15:48:53.221877 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73710567-5e80-472c-a38d-9c6281e5ad10" path="/var/lib/kubelet/pods/73710567-5e80-472c-a38d-9c6281e5ad10/volumes" Nov 28 15:48:53 crc kubenswrapper[4805]: I1128 15:48:53.846145 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ac086fdf-9a1a-40ab-ab2f-8a33dae28290","Type":"ContainerStarted","Data":"fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b"} Nov 28 15:48:53 crc kubenswrapper[4805]: I1128 15:48:53.881508 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.8814766179999998 podStartE2EDuration="2.881476618s" podCreationTimestamp="2025-11-28 15:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:48:53.861813113 +0000 UTC m=+1360.911604504" watchObservedRunningTime="2025-11-28 15:48:53.881476618 +0000 UTC m=+1360.931267949" Nov 28 15:48:56 crc kubenswrapper[4805]: I1128 15:48:56.504844 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 15:48:56 crc kubenswrapper[4805]: I1128 15:48:56.505467 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 15:48:57 crc kubenswrapper[4805]: I1128 15:48:57.242219 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 15:48:58 crc kubenswrapper[4805]: I1128 15:48:58.500748 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 15:48:58 crc kubenswrapper[4805]: I1128 15:48:58.501073 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 15:48:59 crc kubenswrapper[4805]: I1128 15:48:59.512736 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 15:48:59 crc kubenswrapper[4805]: I1128 15:48:59.513086 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 15:49:01 crc kubenswrapper[4805]: I1128 15:49:01.504549 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 15:49:01 crc kubenswrapper[4805]: I1128 15:49:01.505102 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 15:49:02 crc kubenswrapper[4805]: I1128 15:49:02.242541 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 15:49:02 crc kubenswrapper[4805]: I1128 15:49:02.268518 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 15:49:02 crc kubenswrapper[4805]: I1128 15:49:02.518695 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 15:49:02 crc kubenswrapper[4805]: I1128 15:49:02.518788 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 15:49:02 crc kubenswrapper[4805]: I1128 15:49:02.967274 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 15:49:05 crc kubenswrapper[4805]: I1128 15:49:05.955031 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 15:49:08 crc kubenswrapper[4805]: I1128 15:49:08.513203 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 15:49:08 crc kubenswrapper[4805]: I1128 15:49:08.514063 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 15:49:08 crc kubenswrapper[4805]: I1128 15:49:08.515727 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 15:49:08 crc kubenswrapper[4805]: I1128 15:49:08.520868 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 15:49:08 crc kubenswrapper[4805]: I1128 15:49:08.986297 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 15:49:08 crc kubenswrapper[4805]: I1128 15:49:08.992965 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 15:49:11 crc kubenswrapper[4805]: I1128 15:49:11.511484 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 15:49:11 crc kubenswrapper[4805]: I1128 15:49:11.512907 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 15:49:11 crc kubenswrapper[4805]: I1128 15:49:11.526773 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 15:49:12 crc kubenswrapper[4805]: I1128 15:49:12.023499 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 15:49:29 crc kubenswrapper[4805]: I1128 15:49:29.922445 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 28 15:49:29 crc kubenswrapper[4805]: I1128 15:49:29.923179 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" containerName="openstackclient" containerID="cri-o://051cad31407ced2cb3d554ef9c21f2839c03af141e5fd2cda0c2a783a4cd32a3" gracePeriod=2 Nov 28 15:49:29 crc kubenswrapper[4805]: I1128 15:49:29.935920 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 28 15:49:29 crc kubenswrapper[4805]: I1128 15:49:29.988184 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7849b77bf-z2qp2"] Nov 28 15:49:29 crc kubenswrapper[4805]: E1128 15:49:29.988571 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" containerName="openstackclient" Nov 28 15:49:29 crc kubenswrapper[4805]: I1128 15:49:29.988587 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" containerName="openstackclient" Nov 28 15:49:29 crc kubenswrapper[4805]: I1128 15:49:29.988773 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" containerName="openstackclient" Nov 28 15:49:29 crc kubenswrapper[4805]: I1128 15:49:29.989734 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.044837 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7849b77bf-z2qp2"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.097372 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.099320 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.127590 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.145988 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95167e93-42b8-4f5f-b5a9-587b9b854f12-logs\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.146079 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.146103 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf5wk\" (UniqueName: \"kubernetes.io/projected/95167e93-42b8-4f5f-b5a9-587b9b854f12-kube-api-access-nf5wk\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.146140 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc55e90-2c22-4329-b3cc-375a68e6fe26-logs\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.146160 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.146182 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hs4m\" (UniqueName: \"kubernetes.io/projected/bbc55e90-2c22-4329-b3cc-375a68e6fe26-kube-api-access-7hs4m\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.146217 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data-custom\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.146240 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-combined-ca-bundle\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.146256 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-combined-ca-bundle\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.146278 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.248922 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf5wk\" (UniqueName: \"kubernetes.io/projected/95167e93-42b8-4f5f-b5a9-587b9b854f12-kube-api-access-nf5wk\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.248997 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc55e90-2c22-4329-b3cc-375a68e6fe26-logs\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.249024 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.249053 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hs4m\" (UniqueName: \"kubernetes.io/projected/bbc55e90-2c22-4329-b3cc-375a68e6fe26-kube-api-access-7hs4m\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.249096 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data-custom\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.249123 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-combined-ca-bundle\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.249139 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-combined-ca-bundle\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.249161 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.249191 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95167e93-42b8-4f5f-b5a9-587b9b854f12-logs\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.249256 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.252408 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc55e90-2c22-4329-b3cc-375a68e6fe26-logs\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.278018 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.289734 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.292613 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95167e93-42b8-4f5f-b5a9-587b9b854f12-logs\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.293183 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data-custom\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.331040 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-combined-ca-bundle\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.331623 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.332492 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-combined-ca-bundle\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.360912 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf5wk\" (UniqueName: \"kubernetes.io/projected/95167e93-42b8-4f5f-b5a9-587b9b854f12-kube-api-access-nf5wk\") pod \"barbican-worker-7849b77bf-z2qp2\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.378865 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hs4m\" (UniqueName: \"kubernetes.io/projected/bbc55e90-2c22-4329-b3cc-375a68e6fe26-kube-api-access-7hs4m\") pod \"barbican-keystone-listener-7bfdbcc6c8-2pxb2\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.417605 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.526675 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-56786c7f4d-vg252"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.528209 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.598506 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-56786c7f4d-vg252"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.618735 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.668500 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-internal-tls-certs\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.668870 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-combined-ca-bundle\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.668934 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.668963 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk5tw\" (UniqueName: \"kubernetes.io/projected/659e741d-dc3f-4463-b368-075d37a3ef0f-kube-api-access-qk5tw\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.669021 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659e741d-dc3f-4463-b368-075d37a3ef0f-logs\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.669052 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data-custom\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.669093 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-public-tls-certs\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.772900 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.772944 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk5tw\" (UniqueName: \"kubernetes.io/projected/659e741d-dc3f-4463-b368-075d37a3ef0f-kube-api-access-qk5tw\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.773008 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659e741d-dc3f-4463-b368-075d37a3ef0f-logs\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.773031 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data-custom\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.773066 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-public-tls-certs\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.773140 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-internal-tls-certs\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.773167 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-combined-ca-bundle\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.785893 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659e741d-dc3f-4463-b368-075d37a3ef0f-logs\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.786765 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-combined-ca-bundle\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.798752 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-public-tls-certs\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.800302 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.805952 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-internal-tls-certs\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.822989 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data-custom\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.833516 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.855435 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.856158 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerName="openstack-network-exporter" containerID="cri-o://c06570b4467d05bf08c564a535eded9242a1ad77a460690c746c25af47263da6" gracePeriod=300 Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.861585 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance73a6-account-delete-qjlt7"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.865022 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk5tw\" (UniqueName: \"kubernetes.io/projected/659e741d-dc3f-4463-b368-075d37a3ef0f-kube-api-access-qk5tw\") pod \"barbican-api-56786c7f4d-vg252\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.886968 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.892499 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.914452 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance73a6-account-delete-qjlt7"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.929525 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder72ad-account-delete-wnr85"] Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.932325 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.978485 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9mb4\" (UniqueName: \"kubernetes.io/projected/cf2e5369-a11b-4150-a291-4aeab9724d82-kube-api-access-r9mb4\") pod \"glance73a6-account-delete-qjlt7\" (UID: \"cf2e5369-a11b-4150-a291-4aeab9724d82\") " pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:30 crc kubenswrapper[4805]: I1128 15:49:30.978623 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf2e5369-a11b-4150-a291-4aeab9724d82-operator-scripts\") pod \"glance73a6-account-delete-qjlt7\" (UID: \"cf2e5369-a11b-4150-a291-4aeab9724d82\") " pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:30 crc kubenswrapper[4805]: E1128 15:49:30.980639 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 15:49:30 crc kubenswrapper[4805]: E1128 15:49:30.980681 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data podName:4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a nodeName:}" failed. No retries permitted until 2025-11-28 15:49:31.480666196 +0000 UTC m=+1398.530457507 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data") pod "rabbitmq-server-0" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a") : configmap "rabbitmq-config-data" not found Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.036875 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.037116 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="624edf72-de15-4026-812b-36d993917176" containerName="ovn-northd" containerID="cri-o://10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" gracePeriod=30 Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.037225 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="624edf72-de15-4026-812b-36d993917176" containerName="openstack-network-exporter" containerID="cri-o://fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574" gracePeriod=30 Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.055126 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementbd4b-account-delete-7bztm"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.056424 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.072437 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder72ad-account-delete-wnr85"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.082437 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf2e5369-a11b-4150-a291-4aeab9724d82-operator-scripts\") pod \"glance73a6-account-delete-qjlt7\" (UID: \"cf2e5369-a11b-4150-a291-4aeab9724d82\") " pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.082539 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd9hs\" (UniqueName: \"kubernetes.io/projected/f912081e-175b-410c-bbfa-daa3a71e2179-kube-api-access-xd9hs\") pod \"cinder72ad-account-delete-wnr85\" (UID: \"f912081e-175b-410c-bbfa-daa3a71e2179\") " pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.082603 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9mb4\" (UniqueName: \"kubernetes.io/projected/cf2e5369-a11b-4150-a291-4aeab9724d82-kube-api-access-r9mb4\") pod \"glance73a6-account-delete-qjlt7\" (UID: \"cf2e5369-a11b-4150-a291-4aeab9724d82\") " pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.082669 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f912081e-175b-410c-bbfa-daa3a71e2179-operator-scripts\") pod \"cinder72ad-account-delete-wnr85\" (UID: \"f912081e-175b-410c-bbfa-daa3a71e2179\") " pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.083424 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf2e5369-a11b-4150-a291-4aeab9724d82-operator-scripts\") pod \"glance73a6-account-delete-qjlt7\" (UID: \"cf2e5369-a11b-4150-a291-4aeab9724d82\") " pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.114424 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementbd4b-account-delete-7bztm"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.133567 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9mb4\" (UniqueName: \"kubernetes.io/projected/cf2e5369-a11b-4150-a291-4aeab9724d82-kube-api-access-r9mb4\") pod \"glance73a6-account-delete-qjlt7\" (UID: \"cf2e5369-a11b-4150-a291-4aeab9724d82\") " pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.196689 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db4967f-554e-4336-913a-fcebe420cf0e-operator-scripts\") pod \"placementbd4b-account-delete-7bztm\" (UID: \"1db4967f-554e-4336-913a-fcebe420cf0e\") " pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.197253 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd9hs\" (UniqueName: \"kubernetes.io/projected/f912081e-175b-410c-bbfa-daa3a71e2179-kube-api-access-xd9hs\") pod \"cinder72ad-account-delete-wnr85\" (UID: \"f912081e-175b-410c-bbfa-daa3a71e2179\") " pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.197381 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmfd7\" (UniqueName: \"kubernetes.io/projected/1db4967f-554e-4336-913a-fcebe420cf0e-kube-api-access-qmfd7\") pod \"placementbd4b-account-delete-7bztm\" (UID: \"1db4967f-554e-4336-913a-fcebe420cf0e\") " pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.197582 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f912081e-175b-410c-bbfa-daa3a71e2179-operator-scripts\") pod \"cinder72ad-account-delete-wnr85\" (UID: \"f912081e-175b-410c-bbfa-daa3a71e2179\") " pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.198420 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f912081e-175b-410c-bbfa-daa3a71e2179-operator-scripts\") pod \"cinder72ad-account-delete-wnr85\" (UID: \"f912081e-175b-410c-bbfa-daa3a71e2179\") " pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.258076 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd9hs\" (UniqueName: \"kubernetes.io/projected/f912081e-175b-410c-bbfa-daa3a71e2179-kube-api-access-xd9hs\") pod \"cinder72ad-account-delete-wnr85\" (UID: \"f912081e-175b-410c-bbfa-daa3a71e2179\") " pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.273991 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.277753 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-zwp6n"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.302346 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db4967f-554e-4336-913a-fcebe420cf0e-operator-scripts\") pod \"placementbd4b-account-delete-7bztm\" (UID: \"1db4967f-554e-4336-913a-fcebe420cf0e\") " pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.302468 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmfd7\" (UniqueName: \"kubernetes.io/projected/1db4967f-554e-4336-913a-fcebe420cf0e-kube-api-access-qmfd7\") pod \"placementbd4b-account-delete-7bztm\" (UID: \"1db4967f-554e-4336-913a-fcebe420cf0e\") " pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.304034 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db4967f-554e-4336-913a-fcebe420cf0e-operator-scripts\") pod \"placementbd4b-account-delete-7bztm\" (UID: \"1db4967f-554e-4336-913a-fcebe420cf0e\") " pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.309859 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.358785 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-g4grc"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.362318 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmfd7\" (UniqueName: \"kubernetes.io/projected/1db4967f-554e-4336-913a-fcebe420cf0e-kube-api-access-qmfd7\") pod \"placementbd4b-account-delete-7bztm\" (UID: \"1db4967f-554e-4336-913a-fcebe420cf0e\") " pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.376956 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerName="ovsdbserver-nb" containerID="cri-o://74986ddf8b72c947528126972ddc4887c37fae3cb83972e4872342f67b61a129" gracePeriod=300 Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.392045 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-zwp6n"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.438896 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.492914 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-g4grc"] Nov 28 15:49:31 crc kubenswrapper[4805]: E1128 15:49:31.510001 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 15:49:31 crc kubenswrapper[4805]: E1128 15:49:31.510062 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data podName:4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a nodeName:}" failed. No retries permitted until 2025-11-28 15:49:32.510047171 +0000 UTC m=+1399.559838482 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data") pod "rabbitmq-server-0" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a") : configmap "rabbitmq-config-data" not found Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.544643 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-hbmtq"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.583783 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-hbmtq"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.620944 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron1f9e-account-delete-vklns"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.622149 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.689545 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-89z6k"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.689790 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-89z6k" podUID="6aff7621-5632-499d-955d-7ea5b9915b54" containerName="openstack-network-exporter" containerID="cri-o://0a2988e5d6f726fd2171f644ba568d97427b00e8ccb737f58ee5c66888ea03f4" gracePeriod=30 Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.720450 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9f5677-6d82-4ee0-bc2e-74c1671be521-operator-scripts\") pod \"neutron1f9e-account-delete-vklns\" (UID: \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\") " pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.720523 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg7wp\" (UniqueName: \"kubernetes.io/projected/1f9f5677-6d82-4ee0-bc2e-74c1671be521-kube-api-access-qg7wp\") pod \"neutron1f9e-account-delete-vklns\" (UID: \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\") " pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.733285 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron1f9e-account-delete-vklns"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.772469 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4q9c2"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.829265 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-vssvf"] Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.839973 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9f5677-6d82-4ee0-bc2e-74c1671be521-operator-scripts\") pod \"neutron1f9e-account-delete-vklns\" (UID: \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\") " pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.840120 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg7wp\" (UniqueName: \"kubernetes.io/projected/1f9f5677-6d82-4ee0-bc2e-74c1671be521-kube-api-access-qg7wp\") pod \"neutron1f9e-account-delete-vklns\" (UID: \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\") " pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.841907 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9f5677-6d82-4ee0-bc2e-74c1671be521-operator-scripts\") pod \"neutron1f9e-account-delete-vklns\" (UID: \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\") " pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:31 crc kubenswrapper[4805]: I1128 15:49:31.900478 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg7wp\" (UniqueName: \"kubernetes.io/projected/1f9f5677-6d82-4ee0-bc2e-74c1671be521-kube-api-access-qg7wp\") pod \"neutron1f9e-account-delete-vklns\" (UID: \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\") " pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.032959 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.033844 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerName="openstack-network-exporter" containerID="cri-o://2d959c7b97ea239b52290b670e32647011147e73ba303def59e8e383702025e0" gracePeriod=300 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.058247 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.076413 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-7f2hz"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.086207 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-7f2hz"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.099886 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.100184 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" podUID="b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" containerName="dnsmasq-dns" containerID="cri-o://f512ffb73a4fae0efab8bd72a241efe3ae50f7565449d1f938e5a9c88c925c05" gracePeriod=10 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.103325 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-zr49k"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.119389 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-zr49k"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.133960 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.135907 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerName="ovsdbserver-sb" containerID="cri-o://47d6ecbcb66a1126a1aefe7cd1c4f66bc52307aa387bd35f8ee20bce900a2305" gracePeriod=300 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.153578 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-kfvps"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.163744 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-kfvps"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.207146 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican8232-account-delete-52zhn"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.209506 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.231938 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican8232-account-delete-52zhn"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.254975 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-bllqv"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.270146 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnz54\" (UniqueName: \"kubernetes.io/projected/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-kube-api-access-gnz54\") pod \"barbican8232-account-delete-52zhn\" (UID: \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\") " pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.270303 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts\") pod \"barbican8232-account-delete-52zhn\" (UID: \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\") " pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:49:32 crc kubenswrapper[4805]: E1128 15:49:32.271164 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:32 crc kubenswrapper[4805]: E1128 15:49:32.271208 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data podName:fb122aae-0e09-46b2-926c-037d25e79477 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:32.771194905 +0000 UTC m=+1399.820986216 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb122aae-0e09-46b2-926c-037d25e79477") : configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.278306 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-bllqv"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.329553 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-2sh5w"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.373819 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts\") pod \"barbican8232-account-delete-52zhn\" (UID: \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\") " pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.378917 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts\") pod \"barbican8232-account-delete-52zhn\" (UID: \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\") " pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.382133 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnz54\" (UniqueName: \"kubernetes.io/projected/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-kube-api-access-gnz54\") pod \"barbican8232-account-delete-52zhn\" (UID: \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\") " pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.383053 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bee11cf6-ec1b-40bb-8030-044d8f06591e/ovsdbserver-nb/0.log" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.383195 4805 generic.go:334] "Generic (PLEG): container finished" podID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerID="c06570b4467d05bf08c564a535eded9242a1ad77a460690c746c25af47263da6" exitCode=2 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.383211 4805 generic.go:334] "Generic (PLEG): container finished" podID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerID="74986ddf8b72c947528126972ddc4887c37fae3cb83972e4872342f67b61a129" exitCode=143 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.383274 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bee11cf6-ec1b-40bb-8030-044d8f06591e","Type":"ContainerDied","Data":"c06570b4467d05bf08c564a535eded9242a1ad77a460690c746c25af47263da6"} Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.383299 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bee11cf6-ec1b-40bb-8030-044d8f06591e","Type":"ContainerDied","Data":"74986ddf8b72c947528126972ddc4887c37fae3cb83972e4872342f67b61a129"} Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.387519 4805 generic.go:334] "Generic (PLEG): container finished" podID="bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" containerID="051cad31407ced2cb3d554ef9c21f2839c03af141e5fd2cda0c2a783a4cd32a3" exitCode=137 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.389667 4805 generic.go:334] "Generic (PLEG): container finished" podID="b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" containerID="f512ffb73a4fae0efab8bd72a241efe3ae50f7565449d1f938e5a9c88c925c05" exitCode=0 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.389709 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" event={"ID":"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad","Type":"ContainerDied","Data":"f512ffb73a4fae0efab8bd72a241efe3ae50f7565449d1f938e5a9c88c925c05"} Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.392411 4805 generic.go:334] "Generic (PLEG): container finished" podID="624edf72-de15-4026-812b-36d993917176" containerID="fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574" exitCode=2 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.392478 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"624edf72-de15-4026-812b-36d993917176","Type":"ContainerDied","Data":"fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574"} Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.401503 4805 generic.go:334] "Generic (PLEG): container finished" podID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerID="2d959c7b97ea239b52290b670e32647011147e73ba303def59e8e383702025e0" exitCode=2 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.401836 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd","Type":"ContainerDied","Data":"2d959c7b97ea239b52290b670e32647011147e73ba303def59e8e383702025e0"} Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.408804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnz54\" (UniqueName: \"kubernetes.io/projected/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-kube-api-access-gnz54\") pod \"barbican8232-account-delete-52zhn\" (UID: \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\") " pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.413820 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-89z6k_6aff7621-5632-499d-955d-7ea5b9915b54/openstack-network-exporter/0.log" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.413883 4805 generic.go:334] "Generic (PLEG): container finished" podID="6aff7621-5632-499d-955d-7ea5b9915b54" containerID="0a2988e5d6f726fd2171f644ba568d97427b00e8ccb737f58ee5c66888ea03f4" exitCode=2 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.413976 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-89z6k" event={"ID":"6aff7621-5632-499d-955d-7ea5b9915b54","Type":"ContainerDied","Data":"0a2988e5d6f726fd2171f644ba568d97427b00e8ccb737f58ee5c66888ea03f4"} Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.416528 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" event={"ID":"bbc55e90-2c22-4329-b3cc-375a68e6fe26","Type":"ContainerStarted","Data":"589d702195e92daa9f00af6a62686aa77c68b1a505d4b127244e5ebcb17036cd"} Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.416580 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" event={"ID":"bbc55e90-2c22-4329-b3cc-375a68e6fe26","Type":"ContainerStarted","Data":"11c1ce48e069b9a3f885a452ae8ee6f1913b07862dfbb3a191d47dc83ecebfac"} Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.419574 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-2sh5w"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.481426 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.539922 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapic2b1-account-delete-wfngd"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.541522 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.555351 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556122 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-server" containerID="cri-o://cbabf490d47a58ab2c8c9f6eacb4cf048920fd4314a573b001947462a0de16b4" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556274 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="swift-recon-cron" containerID="cri-o://7413e61bbe43f8d00d9ac4ea2485eca35952ef72195b1eb40171e6705175ffcd" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556388 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="rsync" containerID="cri-o://841feae0d3f809ed9e79dff8941de56c70aafb0de8f991cd7453d48a9c67cc4c" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556448 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-replicator" containerID="cri-o://10fe5f001457130421ee90b15666db980a7890c30af98e597a996d50a50c4e9a" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556494 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-server" containerID="cri-o://831f948c09940fb083fcf963fed4d7066d04ebc8f2adc914bc711740e0fcdf49" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556468 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-auditor" containerID="cri-o://7c0cf000e71a5c9d2c9ce21d71ecb228ec004603af520ea9f4b7232773601e18" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556536 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-reaper" containerID="cri-o://663742bde9628596b637c938ce19b2d482d18491b8629d57987967e1b1b24891" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556589 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-auditor" containerID="cri-o://d259dddfea43ce8231ff4ea937ce34f8e372f66bcd4c5effa7aec7ce0ffa8390" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556598 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-expirer" containerID="cri-o://7bc27ced72f3c03367ba7ebd0e3bbe80199cea1ba0a0d9701c61d76ee2259b8b" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556635 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-replicator" containerID="cri-o://3ead42c6ea605c5d27888381393dc51bf09a9814de73cd305dbca46329a8a17d" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556642 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-updater" containerID="cri-o://68a336481b5b62ed5a8243f246583d3e81d48c711638463168818d7ae197b2fe" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556699 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-auditor" containerID="cri-o://4511ff31bb1feea8dab2de6f562fa4e27fc7ba92977a8fbc6bf26f4714ba0be0" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556738 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-replicator" containerID="cri-o://7968f4c1d7d9d723379b1a1881a1e60cc6af6ce686c094d24597209d9d9c45cd" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556770 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-server" containerID="cri-o://a92aa0f7330344e2f61dbf42e3f768648b41c7554cad6173713ec787c88993fc" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.556801 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-updater" containerID="cri-o://fbaa7d95c8f6e7a6d621aa209aea65439fd1c81e062456fd47f282542a39aea3" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.575427 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic2b1-account-delete-wfngd"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.585606 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d7a9256-533a-4985-9a62-9bbe76af9e54-operator-scripts\") pod \"novaapic2b1-account-delete-wfngd\" (UID: \"9d7a9256-533a-4985-9a62-9bbe76af9e54\") " pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.585756 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svmdh\" (UniqueName: \"kubernetes.io/projected/9d7a9256-533a-4985-9a62-9bbe76af9e54-kube-api-access-svmdh\") pod \"novaapic2b1-account-delete-wfngd\" (UID: \"9d7a9256-533a-4985-9a62-9bbe76af9e54\") " pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:32 crc kubenswrapper[4805]: E1128 15:49:32.586015 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 15:49:32 crc kubenswrapper[4805]: E1128 15:49:32.586102 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data podName:4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a nodeName:}" failed. No retries permitted until 2025-11-28 15:49:34.586085328 +0000 UTC m=+1401.635876639 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data") pod "rabbitmq-server-0" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a") : configmap "rabbitmq-config-data" not found Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.586527 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.586772 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerName="cinder-scheduler" containerID="cri-o://4f7f7afc6966bf70692a1ec631ffcd04f68c6d162a03e988818a686b37ca27b5" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.587174 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerName="probe" containerID="cri-o://8007f0c48cb665204284cbdc1681cef5de09b06e0e0fbddc280c197c6fe89921" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.597444 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-55d54b965d-bfqn6"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.597785 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-55d54b965d-bfqn6" podUID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerName="placement-log" containerID="cri-o://3662b2162435ede3ed240fafeef413deeef6d00aecf000c0f5887352f3c5a777" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.597988 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-55d54b965d-bfqn6" podUID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerName="placement-api" containerID="cri-o://fe684bfa9591b55be6b83497c5f2059493fe4abdfa52867aaec588e8310d8893" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.603446 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell08782-account-delete-phg59"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.605128 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.612573 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.612809 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerName="glance-log" containerID="cri-o://4d900ec2e1183b4ce27e8479bb7bc8fa6d050fc7702be965871220be9e57e825" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.612967 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerName="glance-httpd" containerID="cri-o://868c845c22044f953c99bf70b59a10a7cf406a71b3f9ce6c6262250153d75b30" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.622790 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell08782-account-delete-phg59"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.644424 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.644728 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api-log" containerID="cri-o://16befe4e2598415b16fc7b244b3cb1d7660ff6773e1cc20677bf867cfd4f1442" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.645219 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api" containerID="cri-o://0cc217baf0960656d6ad49571b76347e41bfd409a11a61ef61f68ae6d1b9a1c5" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.654973 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6bdfcdc49-v4hgd"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.655260 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6bdfcdc49-v4hgd" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" containerName="neutron-api" containerID="cri-o://7569e16b1b6d818541ba1419b8d0036700d58b74a89a37e7bc67e8d06bfaeb15" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.655626 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6bdfcdc49-v4hgd" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" containerName="neutron-httpd" containerID="cri-o://c253b5bf7e841bb27ed1346fc3c401ce938ec4f3dce125f741a8913badce2e99" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.668648 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.668889 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerName="glance-log" containerID="cri-o://5290a7c9695a774d3cd36d80cc41c5b6f08e321f90789c87b7e2f50f6f5d2099" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.669207 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerName="glance-httpd" containerID="cri-o://31105a43fdab9daa89f1ca09a13406cb4b8904817a89036a2573227173287ae5" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.683253 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6f8fc6d4f-g6nbp"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.683487 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" podUID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerName="proxy-httpd" containerID="cri-o://f341ac5588a30f34bf1c5785f352e9e8bcbb991ecae2e3a87e2814de74b6c3c7" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.683904 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" podUID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerName="proxy-server" containerID="cri-o://a02a9bd5b077fb94a3c91b016fbc51400e43005c07086b7606cdb1ad65200db2" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.688584 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d7a9256-533a-4985-9a62-9bbe76af9e54-operator-scripts\") pod \"novaapic2b1-account-delete-wfngd\" (UID: \"9d7a9256-533a-4985-9a62-9bbe76af9e54\") " pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.688682 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd6fbaf-5702-415e-8a1d-85d2c4792031-operator-scripts\") pod \"novacell08782-account-delete-phg59\" (UID: \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\") " pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.688710 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svmdh\" (UniqueName: \"kubernetes.io/projected/9d7a9256-533a-4985-9a62-9bbe76af9e54-kube-api-access-svmdh\") pod \"novaapic2b1-account-delete-wfngd\" (UID: \"9d7a9256-533a-4985-9a62-9bbe76af9e54\") " pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.688730 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2s4b\" (UniqueName: \"kubernetes.io/projected/ecd6fbaf-5702-415e-8a1d-85d2c4792031-kube-api-access-w2s4b\") pod \"novacell08782-account-delete-phg59\" (UID: \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\") " pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.689566 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d7a9256-533a-4985-9a62-9bbe76af9e54-operator-scripts\") pod \"novaapic2b1-account-delete-wfngd\" (UID: \"9d7a9256-533a-4985-9a62-9bbe76af9e54\") " pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.704432 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.710371 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.711274 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svmdh\" (UniqueName: \"kubernetes.io/projected/9d7a9256-533a-4985-9a62-9bbe76af9e54-kube-api-access-svmdh\") pod \"novaapic2b1-account-delete-wfngd\" (UID: \"9d7a9256-533a-4985-9a62-9bbe76af9e54\") " pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.722205 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.732943 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.733260 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-log" containerID="cri-o://099d02bc7f77ee26f120af8f3d1407c72c14067b902a1e28ada5fa2095d8e0a1" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.733941 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-metadata" containerID="cri-o://5b7e15d112969a7b953c22fe09274e56dd365f74ce49e846d9443a14feb3ce7a" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.771150 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nv5zc"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.778885 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.779034 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-89z6k_6aff7621-5632-499d-955d-7ea5b9915b54/openstack-network-exporter/0.log" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.779087 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.808641 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd6fbaf-5702-415e-8a1d-85d2c4792031-operator-scripts\") pod \"novacell08782-account-delete-phg59\" (UID: \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\") " pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.808989 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2s4b\" (UniqueName: \"kubernetes.io/projected/ecd6fbaf-5702-415e-8a1d-85d2c4792031-kube-api-access-w2s4b\") pod \"novacell08782-account-delete-phg59\" (UID: \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\") " pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:32 crc kubenswrapper[4805]: E1128 15:49:32.809525 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:32 crc kubenswrapper[4805]: E1128 15:49:32.809616 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data podName:fb122aae-0e09-46b2-926c-037d25e79477 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:33.809583778 +0000 UTC m=+1400.859375089 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb122aae-0e09-46b2-926c-037d25e79477") : configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.811037 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd6fbaf-5702-415e-8a1d-85d2c4792031-operator-scripts\") pod \"novacell08782-account-delete-phg59\" (UID: \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\") " pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.813582 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.843742 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2s4b\" (UniqueName: \"kubernetes.io/projected/ecd6fbaf-5702-415e-8a1d-85d2c4792031-kube-api-access-w2s4b\") pod \"novacell08782-account-delete-phg59\" (UID: \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\") " pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.911654 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovn-rundir\") pod \"6aff7621-5632-499d-955d-7ea5b9915b54\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.911719 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aff7621-5632-499d-955d-7ea5b9915b54-config\") pod \"6aff7621-5632-499d-955d-7ea5b9915b54\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.911811 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-combined-ca-bundle\") pod \"6aff7621-5632-499d-955d-7ea5b9915b54\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.911903 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59pnx\" (UniqueName: \"kubernetes.io/projected/6aff7621-5632-499d-955d-7ea5b9915b54-kube-api-access-59pnx\") pod \"6aff7621-5632-499d-955d-7ea5b9915b54\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.911932 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-metrics-certs-tls-certs\") pod \"6aff7621-5632-499d-955d-7ea5b9915b54\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.912031 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovs-rundir\") pod \"6aff7621-5632-499d-955d-7ea5b9915b54\" (UID: \"6aff7621-5632-499d-955d-7ea5b9915b54\") " Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.924472 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "6aff7621-5632-499d-955d-7ea5b9915b54" (UID: "6aff7621-5632-499d-955d-7ea5b9915b54"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.925060 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aff7621-5632-499d-955d-7ea5b9915b54-config" (OuterVolumeSpecName: "config") pod "6aff7621-5632-499d-955d-7ea5b9915b54" (UID: "6aff7621-5632-499d-955d-7ea5b9915b54"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.937583 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.938417 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "6aff7621-5632-499d-955d-7ea5b9915b54" (UID: "6aff7621-5632-499d-955d-7ea5b9915b54"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.939245 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-catalog-content\") pod \"redhat-operators-nv5zc\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.939315 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-utilities\") pod \"redhat-operators-nv5zc\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.939453 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq79p\" (UniqueName: \"kubernetes.io/projected/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-kube-api-access-cq79p\") pod \"redhat-operators-nv5zc\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.940710 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerName="rabbitmq" containerID="cri-o://cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9" gracePeriod=604800 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.941346 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bee11cf6-ec1b-40bb-8030-044d8f06591e/ovsdbserver-nb/0.log" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.941862 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-log" containerID="cri-o://58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.942457 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.942558 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-api" containerID="cri-o://dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649" gracePeriod=30 Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.943064 4805 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovs-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.943086 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6aff7621-5632-499d-955d-7ea5b9915b54-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.943099 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aff7621-5632-499d-955d-7ea5b9915b54-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:32 crc kubenswrapper[4805]: I1128 15:49:32.951737 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aff7621-5632-499d-955d-7ea5b9915b54-kube-api-access-59pnx" (OuterVolumeSpecName: "kube-api-access-59pnx") pod "6aff7621-5632-499d-955d-7ea5b9915b54" (UID: "6aff7621-5632-499d-955d-7ea5b9915b54"). InnerVolumeSpecName "kube-api-access-59pnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.000380 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nv5zc"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.014610 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.014880 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" podUID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerName="barbican-keystone-listener-log" containerID="cri-o://66fc00b5a0c7d5310a8c3fbb5a766ab2e30e4c1049270d558c221c8c94d6292e" gracePeriod=30 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.015006 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" podUID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerName="barbican-keystone-listener" containerID="cri-o://02772248390f9aa46863e063f9eac9961344ec1704c67c8cfde40db99e473634" gracePeriod=30 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.033312 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-97ccb4d45-ddn2r"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.033584 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-97ccb4d45-ddn2r" podUID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerName="barbican-worker-log" containerID="cri-o://c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5" gracePeriod=30 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.036403 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-97ccb4d45-ddn2r" podUID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerName="barbican-worker" containerID="cri-o://1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de" gracePeriod=30 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.046878 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.047561 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-combined-ca-bundle\") pod \"bee11cf6-ec1b-40bb-8030-044d8f06591e\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.047631 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-scripts\") pod \"bee11cf6-ec1b-40bb-8030-044d8f06591e\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.047702 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwlbk\" (UniqueName: \"kubernetes.io/projected/bee11cf6-ec1b-40bb-8030-044d8f06591e-kube-api-access-fwlbk\") pod \"bee11cf6-ec1b-40bb-8030-044d8f06591e\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.047729 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-metrics-certs-tls-certs\") pod \"bee11cf6-ec1b-40bb-8030-044d8f06591e\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.047768 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"bee11cf6-ec1b-40bb-8030-044d8f06591e\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.047821 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-config\") pod \"bee11cf6-ec1b-40bb-8030-044d8f06591e\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.047892 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdb-rundir\") pod \"bee11cf6-ec1b-40bb-8030-044d8f06591e\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.047919 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdbserver-nb-tls-certs\") pod \"bee11cf6-ec1b-40bb-8030-044d8f06591e\" (UID: \"bee11cf6-ec1b-40bb-8030-044d8f06591e\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.048257 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-catalog-content\") pod \"redhat-operators-nv5zc\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.048278 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-utilities\") pod \"redhat-operators-nv5zc\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.048308 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq79p\" (UniqueName: \"kubernetes.io/projected/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-kube-api-access-cq79p\") pod \"redhat-operators-nv5zc\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.048403 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59pnx\" (UniqueName: \"kubernetes.io/projected/6aff7621-5632-499d-955d-7ea5b9915b54-kube-api-access-59pnx\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.052105 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "bee11cf6-ec1b-40bb-8030-044d8f06591e" (UID: "bee11cf6-ec1b-40bb-8030-044d8f06591e"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.059302 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-catalog-content\") pod \"redhat-operators-nv5zc\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.059648 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-utilities\") pod \"redhat-operators-nv5zc\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.063513 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-scripts" (OuterVolumeSpecName: "scripts") pod "bee11cf6-ec1b-40bb-8030-044d8f06591e" (UID: "bee11cf6-ec1b-40bb-8030-044d8f06591e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.073614 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-config" (OuterVolumeSpecName: "config") pod "bee11cf6-ec1b-40bb-8030-044d8f06591e" (UID: "bee11cf6-ec1b-40bb-8030-044d8f06591e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.084790 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.097235 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq79p\" (UniqueName: \"kubernetes.io/projected/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-kube-api-access-cq79p\") pod \"redhat-operators-nv5zc\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.112408 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.113231 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e50653a2-4881-42fd-bc8a-35e5f8e2f48d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73" gracePeriod=30 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.149497 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-56786c7f4d-vg252"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.160620 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.175180 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7849b77bf-z2qp2"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.175556 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "bee11cf6-ec1b-40bb-8030-044d8f06591e" (UID: "bee11cf6-ec1b-40bb-8030-044d8f06591e"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.177081 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.177107 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.177118 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bee11cf6-ec1b-40bb-8030-044d8f06591e-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.177140 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.182167 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bee11cf6-ec1b-40bb-8030-044d8f06591e-kube-api-access-fwlbk" (OuterVolumeSpecName: "kube-api-access-fwlbk") pod "bee11cf6-ec1b-40bb-8030-044d8f06591e" (UID: "bee11cf6-ec1b-40bb-8030-044d8f06591e"). InnerVolumeSpecName "kube-api-access-fwlbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.215333 4805 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" secret="" err="secret \"barbican-barbican-dockercfg-gzh2w\" not found" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.230488 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6aff7621-5632-499d-955d-7ea5b9915b54" (UID: "6aff7621-5632-499d-955d-7ea5b9915b54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.234928 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="300c24d4-04ee-4913-8e98-954c3f5c6f94" path="/var/lib/kubelet/pods/300c24d4-04ee-4913-8e98-954c3f5c6f94/volumes" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.236085 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37bb6d65-078d-4d19-bc10-dc50ea71edae" path="/var/lib/kubelet/pods/37bb6d65-078d-4d19-bc10-dc50ea71edae/volumes" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.237947 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b75302a-7cbf-4b5a-ad36-f6cdabe27221" path="/var/lib/kubelet/pods/6b75302a-7cbf-4b5a-ad36-f6cdabe27221/volumes" Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.238062 4805 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.196:40732->38.102.83.196:46305: write tcp 38.102.83.196:40732->38.102.83.196:46305: write: broken pipe Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.239221 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a64cdf5b-5d0f-431f-a96f-ec0351559d2f" path="/var/lib/kubelet/pods/a64cdf5b-5d0f-431f-a96f-ec0351559d2f/volumes" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.240621 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2f218fa-9d63-46f4-ba0d-3b9166d23672" path="/var/lib/kubelet/pods/b2f218fa-9d63-46f4-ba0d-3b9166d23672/volumes" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.242255 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d86dfa-8c46-40bb-8155-438f6d5e06e6" path="/var/lib/kubelet/pods/b6d86dfa-8c46-40bb-8155-438f6d5e06e6/volumes" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.243223 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e013d7bd-1c72-4b87-be4b-83f895797124" path="/var/lib/kubelet/pods/e013d7bd-1c72-4b87-be4b-83f895797124/volumes" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.245008 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e63a3a44-329f-4d06-af10-a8ac7c72dcc4" path="/var/lib/kubelet/pods/e63a3a44-329f-4d06-af10-a8ac7c72dcc4/volumes" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.249390 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.257941 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5fc4b9b6bd-9nsgk"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.257996 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-wn5s2"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.266509 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-wn5s2"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.268402 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bee11cf6-ec1b-40bb-8030-044d8f06591e" (UID: "bee11cf6-ec1b-40bb-8030-044d8f06591e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.274785 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-1502-account-create-update-2sqhk"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.278865 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.278890 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.278900 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.278909 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwlbk\" (UniqueName: \"kubernetes.io/projected/bee11cf6-ec1b-40bb-8030-044d8f06591e-kube-api-access-fwlbk\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.279421 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-1502-account-create-update-2sqhk"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.289414 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cfqg8"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.380899 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.381121 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" containerName="nova-cell0-conductor-conductor" containerID="cri-o://3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339" gracePeriod=30 Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.382485 4805 secret.go:188] Couldn't get secret openstack/barbican-api-config-data: secret "barbican-api-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.382541 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom podName:5c4171ee-19b0-44c4-8e98-d14e11b9e1a6 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:33.882527111 +0000 UTC m=+1400.932318422 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom") pod "barbican-api-5fc4b9b6bd-9nsgk" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6") : secret "barbican-api-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.382580 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.382601 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data podName:5c4171ee-19b0-44c4-8e98-d14e11b9e1a6 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:33.882595693 +0000 UTC m=+1400.932387004 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data") pod "barbican-api-5fc4b9b6bd-9nsgk" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6") : secret "barbican-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.403843 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cfqg8"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.408876 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "bee11cf6-ec1b-40bb-8030-044d8f06591e" (UID: "bee11cf6-ec1b-40bb-8030-044d8f06591e"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.416878 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.417597 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="51493fc6-cc57-4bbf-a6a8-0864160ade3f" containerName="nova-cell1-conductor-conductor" containerID="cri-o://273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1" gracePeriod=30 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.429627 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "6aff7621-5632-499d-955d-7ea5b9915b54" (UID: "6aff7621-5632-499d-955d-7ea5b9915b54"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.444767 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pc5z4"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.455469 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pc5z4"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.464976 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.465224 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ac086fdf-9a1a-40ab-ab2f-8a33dae28290" containerName="nova-scheduler-scheduler" containerID="cri-o://fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b" gracePeriod=30 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.477052 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "bee11cf6-ec1b-40bb-8030-044d8f06591e" (UID: "bee11cf6-ec1b-40bb-8030-044d8f06591e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.485615 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.485651 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aff7621-5632-499d-955d-7ea5b9915b54-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.485662 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bee11cf6-ec1b-40bb-8030-044d8f06591e-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.501692 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" containerID="cri-o://e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" gracePeriod=29 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.522073 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="452c347f-4ee3-46de-ba8f-c83300966f5d" containerName="galera" containerID="cri-o://8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f" gracePeriod=30 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.543139 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.545225 4805 generic.go:334] "Generic (PLEG): container finished" podID="d7348256-32de-4696-9637-96d7323831ed" containerID="099d02bc7f77ee26f120af8f3d1407c72c14067b902a1e28ada5fa2095d8e0a1" exitCode=143 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.546020 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7348256-32de-4696-9637-96d7323831ed","Type":"ContainerDied","Data":"099d02bc7f77ee26f120af8f3d1407c72c14067b902a1e28ada5fa2095d8e0a1"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.564725 4805 generic.go:334] "Generic (PLEG): container finished" podID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerID="5290a7c9695a774d3cd36d80cc41c5b6f08e321f90789c87b7e2f50f6f5d2099" exitCode=143 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.565034 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-56786c7f4d-vg252"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.565062 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c021413-239d-4ba0-979d-1c4b25d5093c","Type":"ContainerDied","Data":"5290a7c9695a774d3cd36d80cc41c5b6f08e321f90789c87b7e2f50f6f5d2099"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.570827 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7849b77bf-z2qp2"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.586742 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder72ad-account-delete-wnr85"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.594650 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance73a6-account-delete-qjlt7"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.599871 4805 generic.go:334] "Generic (PLEG): container finished" podID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerID="66fc00b5a0c7d5310a8c3fbb5a766ab2e30e4c1049270d558c221c8c94d6292e" exitCode=143 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.599972 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" event={"ID":"a363e2d1-0a91-4578-b9e4-b736b7931b03","Type":"ContainerDied","Data":"66fc00b5a0c7d5310a8c3fbb5a766ab2e30e4c1049270d558c221c8c94d6292e"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.601346 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementbd4b-account-delete-7bztm"] Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.655095 4805 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 28 15:49:33 crc kubenswrapper[4805]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 28 15:49:33 crc kubenswrapper[4805]: + source /usr/local/bin/container-scripts/functions Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNBridge=br-int Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNRemote=tcp:localhost:6642 Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNEncapType=geneve Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNAvailabilityZones= Nov 28 15:49:33 crc kubenswrapper[4805]: ++ EnableChassisAsGateway=true Nov 28 15:49:33 crc kubenswrapper[4805]: ++ PhysicalNetworks= Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNHostName= Nov 28 15:49:33 crc kubenswrapper[4805]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 28 15:49:33 crc kubenswrapper[4805]: ++ ovs_dir=/var/lib/openvswitch Nov 28 15:49:33 crc kubenswrapper[4805]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 28 15:49:33 crc kubenswrapper[4805]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 28 15:49:33 crc kubenswrapper[4805]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + sleep 0.5 Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + sleep 0.5 Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + sleep 0.5 Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + cleanup_ovsdb_server_semaphore Nov 28 15:49:33 crc kubenswrapper[4805]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 15:49:33 crc kubenswrapper[4805]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 28 15:49:33 crc kubenswrapper[4805]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-vssvf" message=< Nov 28 15:49:33 crc kubenswrapper[4805]: Exiting ovsdb-server (5) [ OK ] Nov 28 15:49:33 crc kubenswrapper[4805]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 28 15:49:33 crc kubenswrapper[4805]: + source /usr/local/bin/container-scripts/functions Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNBridge=br-int Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNRemote=tcp:localhost:6642 Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNEncapType=geneve Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNAvailabilityZones= Nov 28 15:49:33 crc kubenswrapper[4805]: ++ EnableChassisAsGateway=true Nov 28 15:49:33 crc kubenswrapper[4805]: ++ PhysicalNetworks= Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNHostName= Nov 28 15:49:33 crc kubenswrapper[4805]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 28 15:49:33 crc kubenswrapper[4805]: ++ ovs_dir=/var/lib/openvswitch Nov 28 15:49:33 crc kubenswrapper[4805]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 28 15:49:33 crc kubenswrapper[4805]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 28 15:49:33 crc kubenswrapper[4805]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + sleep 0.5 Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + sleep 0.5 Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + sleep 0.5 Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + cleanup_ovsdb_server_semaphore Nov 28 15:49:33 crc kubenswrapper[4805]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 15:49:33 crc kubenswrapper[4805]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 28 15:49:33 crc kubenswrapper[4805]: > Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.655138 4805 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 28 15:49:33 crc kubenswrapper[4805]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 28 15:49:33 crc kubenswrapper[4805]: + source /usr/local/bin/container-scripts/functions Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNBridge=br-int Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNRemote=tcp:localhost:6642 Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNEncapType=geneve Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNAvailabilityZones= Nov 28 15:49:33 crc kubenswrapper[4805]: ++ EnableChassisAsGateway=true Nov 28 15:49:33 crc kubenswrapper[4805]: ++ PhysicalNetworks= Nov 28 15:49:33 crc kubenswrapper[4805]: ++ OVNHostName= Nov 28 15:49:33 crc kubenswrapper[4805]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 28 15:49:33 crc kubenswrapper[4805]: ++ ovs_dir=/var/lib/openvswitch Nov 28 15:49:33 crc kubenswrapper[4805]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 28 15:49:33 crc kubenswrapper[4805]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 28 15:49:33 crc kubenswrapper[4805]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + sleep 0.5 Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + sleep 0.5 Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + sleep 0.5 Nov 28 15:49:33 crc kubenswrapper[4805]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 28 15:49:33 crc kubenswrapper[4805]: + cleanup_ovsdb_server_semaphore Nov 28 15:49:33 crc kubenswrapper[4805]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 28 15:49:33 crc kubenswrapper[4805]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 28 15:49:33 crc kubenswrapper[4805]: > pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" containerID="cri-o://2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.655165 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" containerID="cri-o://2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" gracePeriod=29 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.667613 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron1f9e-account-delete-vklns"] Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691210 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="841feae0d3f809ed9e79dff8941de56c70aafb0de8f991cd7453d48a9c67cc4c" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691525 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="7bc27ced72f3c03367ba7ebd0e3bbe80199cea1ba0a0d9701c61d76ee2259b8b" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691534 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="68a336481b5b62ed5a8243f246583d3e81d48c711638463168818d7ae197b2fe" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691542 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="4511ff31bb1feea8dab2de6f562fa4e27fc7ba92977a8fbc6bf26f4714ba0be0" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691549 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="7968f4c1d7d9d723379b1a1881a1e60cc6af6ce686c094d24597209d9d9c45cd" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691557 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="a92aa0f7330344e2f61dbf42e3f768648b41c7554cad6173713ec787c88993fc" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691563 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="fbaa7d95c8f6e7a6d621aa209aea65439fd1c81e062456fd47f282542a39aea3" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691569 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="7c0cf000e71a5c9d2c9ce21d71ecb228ec004603af520ea9f4b7232773601e18" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691867 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="10fe5f001457130421ee90b15666db980a7890c30af98e597a996d50a50c4e9a" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691881 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="831f948c09940fb083fcf963fed4d7066d04ebc8f2adc914bc711740e0fcdf49" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691888 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="663742bde9628596b637c938ce19b2d482d18491b8629d57987967e1b1b24891" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691894 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="d259dddfea43ce8231ff4ea937ce34f8e372f66bcd4c5effa7aec7ce0ffa8390" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691901 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="3ead42c6ea605c5d27888381393dc51bf09a9814de73cd305dbca46329a8a17d" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.691907 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="cbabf490d47a58ab2c8c9f6eacb4cf048920fd4314a573b001947462a0de16b4" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692015 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"841feae0d3f809ed9e79dff8941de56c70aafb0de8f991cd7453d48a9c67cc4c"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692046 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"7bc27ced72f3c03367ba7ebd0e3bbe80199cea1ba0a0d9701c61d76ee2259b8b"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692060 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"68a336481b5b62ed5a8243f246583d3e81d48c711638463168818d7ae197b2fe"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692070 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"4511ff31bb1feea8dab2de6f562fa4e27fc7ba92977a8fbc6bf26f4714ba0be0"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692079 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"7968f4c1d7d9d723379b1a1881a1e60cc6af6ce686c094d24597209d9d9c45cd"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692087 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"a92aa0f7330344e2f61dbf42e3f768648b41c7554cad6173713ec787c88993fc"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692095 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"fbaa7d95c8f6e7a6d621aa209aea65439fd1c81e062456fd47f282542a39aea3"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692104 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"7c0cf000e71a5c9d2c9ce21d71ecb228ec004603af520ea9f4b7232773601e18"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692112 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"10fe5f001457130421ee90b15666db980a7890c30af98e597a996d50a50c4e9a"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692120 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"831f948c09940fb083fcf963fed4d7066d04ebc8f2adc914bc711740e0fcdf49"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692130 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"663742bde9628596b637c938ce19b2d482d18491b8629d57987967e1b1b24891"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692139 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"d259dddfea43ce8231ff4ea937ce34f8e372f66bcd4c5effa7aec7ce0ffa8390"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692147 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"3ead42c6ea605c5d27888381393dc51bf09a9814de73cd305dbca46329a8a17d"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.692156 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"cbabf490d47a58ab2c8c9f6eacb4cf048920fd4314a573b001947462a0de16b4"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.759586 4805 generic.go:334] "Generic (PLEG): container finished" podID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerID="c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5" exitCode=143 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.759682 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-97ccb4d45-ddn2r" event={"ID":"abfd499c-e83b-4616-a80c-29a7e8e750e2","Type":"ContainerDied","Data":"c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.787044 4805 generic.go:334] "Generic (PLEG): container finished" podID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerID="a02a9bd5b077fb94a3c91b016fbc51400e43005c07086b7606cdb1ad65200db2" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.787072 4805 generic.go:334] "Generic (PLEG): container finished" podID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerID="f341ac5588a30f34bf1c5785f352e9e8bcbb991ecae2e3a87e2814de74b6c3c7" exitCode=0 Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.787106 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" event={"ID":"33489fb0-b55a-4544-8a36-c4e0c9ca10f0","Type":"ContainerDied","Data":"a02a9bd5b077fb94a3c91b016fbc51400e43005c07086b7606cdb1ad65200db2"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.787131 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" event={"ID":"33489fb0-b55a-4544-8a36-c4e0c9ca10f0","Type":"ContainerDied","Data":"f341ac5588a30f34bf1c5785f352e9e8bcbb991ecae2e3a87e2814de74b6c3c7"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.787309 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.863304 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" event={"ID":"bbc55e90-2c22-4329-b3cc-375a68e6fe26","Type":"ContainerStarted","Data":"5ac62896faa25edf067e9eba6f68e421a1b233d5e85b4ee414fe3961dc9ac78d"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.863962 4805 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" secret="" err="secret \"barbican-barbican-dockercfg-gzh2w\" not found" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.865298 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.924712 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-config\") pod \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.924809 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-nb\") pod \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.924834 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-swift-storage-0\") pod \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.932340 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-svc\") pod \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.934138 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqhb7\" (UniqueName: \"kubernetes.io/projected/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-kube-api-access-wqhb7\") pod \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.934379 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-sb\") pod \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\" (UID: \"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad\") " Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.941291 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" podStartSLOduration=3.941271043 podStartE2EDuration="3.941271043s" podCreationTimestamp="2025-11-28 15:49:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:49:33.913197501 +0000 UTC m=+1400.962988822" watchObservedRunningTime="2025-11-28 15:49:33.941271043 +0000 UTC m=+1400.991062354" Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.960679 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.960752 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data podName:fb122aae-0e09-46b2-926c-037d25e79477 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:35.960734298 +0000 UTC m=+1403.010525609 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb122aae-0e09-46b2-926c-037d25e79477") : configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.961135 4805 secret.go:188] Couldn't get secret openstack/barbican-api-config-data: secret "barbican-api-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.961164 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom podName:5c4171ee-19b0-44c4-8e98-d14e11b9e1a6 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:34.961156609 +0000 UTC m=+1402.010947920 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom") pod "barbican-api-5fc4b9b6bd-9nsgk" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6") : secret "barbican-api-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.961207 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: E1128 15:49:33.963908 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data podName:5c4171ee-19b0-44c4-8e98-d14e11b9e1a6 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:34.963824562 +0000 UTC m=+1402.013615873 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data") pod "barbican-api-5fc4b9b6bd-9nsgk" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6") : secret "barbican-config-data" not found Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.966054 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-kube-api-access-wqhb7" (OuterVolumeSpecName: "kube-api-access-wqhb7") pod "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" (UID: "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad"). InnerVolumeSpecName "kube-api-access-wqhb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.974759 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.974951 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_07201b49-2b7c-4e44-a6a1-3a2dec9df8bd/ovsdbserver-sb/0.log" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.975031 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.975087 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bee11cf6-ec1b-40bb-8030-044d8f06591e","Type":"ContainerDied","Data":"fa7415ffdccf3dfa7986ea148ea45e4b64a565d2af7811fce09af66a733bab92"} Nov 28 15:49:33 crc kubenswrapper[4805]: I1128 15:49:33.975122 4805 scope.go:117] "RemoveContainer" containerID="c06570b4467d05bf08c564a535eded9242a1ad77a460690c746c25af47263da6" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.105774 4805 generic.go:334] "Generic (PLEG): container finished" podID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerID="8007f0c48cb665204284cbdc1681cef5de09b06e0e0fbddc280c197c6fe89921" exitCode=0 Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.107336 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54ff5706-18a9-4e71-8289-e3d9e5d9af23","Type":"ContainerDied","Data":"8007f0c48cb665204284cbdc1681cef5de09b06e0e0fbddc280c197c6fe89921"} Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.108519 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-combined-ca-bundle\") pod \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.108596 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdb-rundir\") pod \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.108621 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-metrics-certs-tls-certs\") pod \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.108666 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-config\") pod \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.108712 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.108776 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config-secret\") pod \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.108923 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-scripts\") pod \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.108956 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdbserver-sb-tls-certs\") pod \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.108991 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-combined-ca-bundle\") pod \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.109033 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsnx2\" (UniqueName: \"kubernetes.io/projected/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-kube-api-access-xsnx2\") pod \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\" (UID: \"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.109059 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config\") pod \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.109078 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcggf\" (UniqueName: \"kubernetes.io/projected/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-kube-api-access-gcggf\") pod \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\" (UID: \"bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc\") " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.115302 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" (UID: "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.117396 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-scripts" (OuterVolumeSpecName: "scripts") pod "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" (UID: "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.117517 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-config" (OuterVolumeSpecName: "config") pod "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" (UID: "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.118111 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqhb7\" (UniqueName: \"kubernetes.io/projected/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-kube-api-access-wqhb7\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.118473 4805 secret.go:188] Couldn't get secret openstack/barbican-keystone-listener-config-data: secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.118552 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:34.618526164 +0000 UTC m=+1401.668317475 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.120310 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.120897 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.121778 4805 generic.go:334] "Generic (PLEG): container finished" podID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerID="3662b2162435ede3ed240fafeef413deeef6d00aecf000c0f5887352f3c5a777" exitCode=143 Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.121836 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55d54b965d-bfqn6" event={"ID":"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6","Type":"ContainerDied","Data":"3662b2162435ede3ed240fafeef413deeef6d00aecf000c0f5887352f3c5a777"} Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.121880 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.121900 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.122184 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.122438 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:34.622328878 +0000 UTC m=+1401.672120189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.128576 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.130688 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.147912 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.148035 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.150074 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_07201b49-2b7c-4e44-a6a1-3a2dec9df8bd/ovsdbserver-sb/0.log" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.150140 4805 generic.go:334] "Generic (PLEG): container finished" podID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerID="47d6ecbcb66a1126a1aefe7cd1c4f66bc52307aa387bd35f8ee20bce900a2305" exitCode=143 Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.150307 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.150499 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"07201b49-2b7c-4e44-a6a1-3a2dec9df8bd","Type":"ContainerDied","Data":"47d6ecbcb66a1126a1aefe7cd1c4f66bc52307aa387bd35f8ee20bce900a2305"} Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.170944 4805 scope.go:117] "RemoveContainer" containerID="74986ddf8b72c947528126972ddc4887c37fae3cb83972e4872342f67b61a129" Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.171310 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.185559 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.189283 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.189338 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" containerName="nova-cell0-conductor-conductor" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.199751 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-kube-api-access-gcggf" (OuterVolumeSpecName: "kube-api-access-gcggf") pod "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" (UID: "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc"). InnerVolumeSpecName "kube-api-access-gcggf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.201169 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.203066 4805 generic.go:334] "Generic (PLEG): container finished" podID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerID="58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194" exitCode=143 Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.203161 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ef711de-53e6-4705-8d2f-469b6dc2d4de","Type":"ContainerDied","Data":"58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194"} Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.210771 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-kube-api-access-xsnx2" (OuterVolumeSpecName: "kube-api-access-xsnx2") pod "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" (UID: "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd"). InnerVolumeSpecName "kube-api-access-xsnx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.212771 4805 generic.go:334] "Generic (PLEG): container finished" podID="8724692d-29b7-46c8-b0dd-802e570050a7" containerID="c253b5bf7e841bb27ed1346fc3c401ce938ec4f3dce125f741a8913badce2e99" exitCode=0 Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.212943 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdfcdc49-v4hgd" event={"ID":"8724692d-29b7-46c8-b0dd-802e570050a7","Type":"ContainerDied","Data":"c253b5bf7e841bb27ed1346fc3c401ce938ec4f3dce125f741a8913badce2e99"} Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.222919 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.222941 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.222950 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsnx2\" (UniqueName: \"kubernetes.io/projected/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-kube-api-access-xsnx2\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.222959 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcggf\" (UniqueName: \"kubernetes.io/projected/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-kube-api-access-gcggf\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.222968 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.271611 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.297761 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" (UID: "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.298979 4805 generic.go:334] "Generic (PLEG): container finished" podID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerID="4d900ec2e1183b4ce27e8479bb7bc8fa6d050fc7702be965871220be9e57e825" exitCode=143 Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.299074 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"19ef9cc1-28ab-4014-897c-9679f3d36443","Type":"ContainerDied","Data":"4d900ec2e1183b4ce27e8479bb7bc8fa6d050fc7702be965871220be9e57e825"} Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.346755 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.365643 4805 generic.go:334] "Generic (PLEG): container finished" podID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerID="16befe4e2598415b16fc7b244b3cb1d7660ff6773e1cc20677bf867cfd4f1442" exitCode=143 Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.366555 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" (UID: "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.366616 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be5d0310-fb90-4d73-94f3-57b105a2a408","Type":"ContainerDied","Data":"16befe4e2598415b16fc7b244b3cb1d7660ff6773e1cc20677bf867cfd4f1442"} Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.411730 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.412496 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-config" (OuterVolumeSpecName: "config") pod "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" (UID: "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.441661 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" (UID: "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.449325 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.449381 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.449396 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.455945 4805 scope.go:117] "RemoveContainer" containerID="2d959c7b97ea239b52290b670e32647011147e73ba303def59e8e383702025e0" Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.456075 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.456785 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api-log" containerID="cri-o://e4e089980980ac67dc67f214af92d43970eb148b6a828b8e14ba26fab9d1df86" gracePeriod=30 Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.456891 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-89z6k" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.457335 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-89z6k" event={"ID":"6aff7621-5632-499d-955d-7ea5b9915b54","Type":"ContainerDied","Data":"950cdf1704e1f2d57fa510e2ac4b2b3f220da2b02ccee75987b07b564e026297"} Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.458413 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api" containerID="cri-o://42eba8d003a7a3309a588e3a7f4e6e3ccf9656400a9603ef4fb2b81bd3c138d4" gracePeriod=30 Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.459333 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" (UID: "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.459477 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.459540 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="624edf72-de15-4026-812b-36d993917176" containerName="ovn-northd" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.480050 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" (UID: "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.535817 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" (UID: "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.537277 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.537789 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" (UID: "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.565389 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.565604 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.565615 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.565623 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.565631 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.613832 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.668003 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.668060 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data podName:4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a nodeName:}" failed. No retries permitted until 2025-11-28 15:49:38.668046453 +0000 UTC m=+1405.717837764 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data") pod "rabbitmq-server-0" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a") : configmap "rabbitmq-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.668390 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.668422 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:35.668415273 +0000 UTC m=+1402.718206584 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.668458 4805 secret.go:188] Couldn't get secret openstack/barbican-keystone-listener-config-data: secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.668479 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:35.668472534 +0000 UTC m=+1402.718263845 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.750294 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" (UID: "b36ae016-8ef7-4c1f-a902-9e6f33d9aaad"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.770269 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" (UID: "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.779121 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" (UID: "bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.781497 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" (UID: "07201b49-2b7c-4e44-a6a1-3a2dec9df8bd"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.785695 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.786654 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.786759 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.786839 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.823299 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell08782-account-delete-phg59"] Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.858936 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic2b1-account-delete-wfngd"] Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.876993 4805 scope.go:117] "RemoveContainer" containerID="47d6ecbcb66a1126a1aefe7cd1c4f66bc52307aa387bd35f8ee20bce900a2305" Nov 28 15:49:34 crc kubenswrapper[4805]: I1128 15:49:34.965858 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.997631 4805 secret.go:188] Couldn't get secret openstack/barbican-api-config-data: secret "barbican-api-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.997717 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom podName:5c4171ee-19b0-44c4-8e98-d14e11b9e1a6 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:36.997696949 +0000 UTC m=+1404.047488310 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom") pod "barbican-api-5fc4b9b6bd-9nsgk" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6") : secret "barbican-api-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.997768 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:34 crc kubenswrapper[4805]: E1128 15:49:34.997841 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data podName:5c4171ee-19b0-44c4-8e98-d14e11b9e1a6 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:36.997824053 +0000 UTC m=+1404.047615454 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data") pod "barbican-api-5fc4b9b6bd-9nsgk" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6") : secret "barbican-config-data" not found Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.003523 4805 scope.go:117] "RemoveContainer" containerID="0a2988e5d6f726fd2171f644ba568d97427b00e8ccb737f58ee5c66888ea03f4" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.090382 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.101641 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-etc-swift\") pod \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.101697 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-log-httpd\") pod \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.101762 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-combined-ca-bundle\") pod \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.101832 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-public-tls-certs\") pod \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.101907 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-config-data\") pod \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.102019 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5h8n\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-kube-api-access-v5h8n\") pod \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.102096 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-internal-tls-certs\") pod \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.102426 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-run-httpd\") pod \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\" (UID: \"33489fb0-b55a-4544-8a36-c4e0c9ca10f0\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.105587 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "33489fb0-b55a-4544-8a36-c4e0c9ca10f0" (UID: "33489fb0-b55a-4544-8a36-c4e0c9ca10f0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.105726 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "33489fb0-b55a-4544-8a36-c4e0c9ca10f0" (UID: "33489fb0-b55a-4544-8a36-c4e0c9ca10f0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.176925 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-89z6k"] Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.195549 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-89z6k"] Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.200570 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-kube-api-access-v5h8n" (OuterVolumeSpecName: "kube-api-access-v5h8n") pod "33489fb0-b55a-4544-8a36-c4e0c9ca10f0" (UID: "33489fb0-b55a-4544-8a36-c4e0c9ca10f0"). InnerVolumeSpecName "kube-api-access-v5h8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.207137 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-vencrypt-tls-certs\") pod \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.207242 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-config-data\") pod \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.207327 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-nova-novncproxy-tls-certs\") pod \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.207467 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5tv6\" (UniqueName: \"kubernetes.io/projected/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-kube-api-access-h5tv6\") pod \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.207522 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-combined-ca-bundle\") pod \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\" (UID: \"e50653a2-4881-42fd-bc8a-35e5f8e2f48d\") " Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.207907 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.207917 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.207925 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5h8n\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-kube-api-access-v5h8n\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.216406 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "33489fb0-b55a-4544-8a36-c4e0c9ca10f0" (UID: "33489fb0-b55a-4544-8a36-c4e0c9ca10f0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.247927 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-kube-api-access-h5tv6" (OuterVolumeSpecName: "kube-api-access-h5tv6") pod "e50653a2-4881-42fd-bc8a-35e5f8e2f48d" (UID: "e50653a2-4881-42fd-bc8a-35e5f8e2f48d"). InnerVolumeSpecName "kube-api-access-h5tv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.280805 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e934e2-71c2-40ee-bcc7-bbd04ecbac9c" path="/var/lib/kubelet/pods/35e934e2-71c2-40ee-bcc7-bbd04ecbac9c/volumes" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.281745 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39c6f2be-4e1d-40c1-8871-a008f39d2d49" path="/var/lib/kubelet/pods/39c6f2be-4e1d-40c1-8871-a008f39d2d49/volumes" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.282762 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="547a6f46-a3b2-42b9-9681-80e8bbfaa4e6" path="/var/lib/kubelet/pods/547a6f46-a3b2-42b9-9681-80e8bbfaa4e6/volumes" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.294864 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aff7621-5632-499d-955d-7ea5b9915b54" path="/var/lib/kubelet/pods/6aff7621-5632-499d-955d-7ea5b9915b54/volumes" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.295708 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79634ed6-4d3f-4488-8123-bce2cb65249d" path="/var/lib/kubelet/pods/79634ed6-4d3f-4488-8123-bce2cb65249d/volumes" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.296336 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc" path="/var/lib/kubelet/pods/bb8cd2a8-99cd-47d2-ba18-c34a9b86abbc/volumes" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.299235 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bee11cf6-ec1b-40bb-8030-044d8f06591e" path="/var/lib/kubelet/pods/bee11cf6-ec1b-40bb-8030-044d8f06591e/volumes" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.312045 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5tv6\" (UniqueName: \"kubernetes.io/projected/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-kube-api-access-h5tv6\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.312294 4805 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.480548 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.488108 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.493609 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.511134 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e50653a2-4881-42fd-bc8a-35e5f8e2f48d" (UID: "e50653a2-4881-42fd-bc8a-35e5f8e2f48d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.518138 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.521533 4805 generic.go:334] "Generic (PLEG): container finished" podID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" exitCode=0 Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.537530 4805 generic.go:334] "Generic (PLEG): container finished" podID="e50653a2-4881-42fd-bc8a-35e5f8e2f48d" containerID="7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73" exitCode=0 Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.537638 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.542985 4805 generic.go:334] "Generic (PLEG): container finished" podID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerID="e4e089980980ac67dc67f214af92d43970eb148b6a828b8e14ba26fab9d1df86" exitCode=143 Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.549201 4805 generic.go:334] "Generic (PLEG): container finished" podID="452c347f-4ee3-46de-ba8f-c83300966f5d" containerID="8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f" exitCode=0 Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.555490 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" podUID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerName="barbican-keystone-listener-log" containerID="cri-o://589d702195e92daa9f00af6a62686aa77c68b1a505d4b127244e5ebcb17036cd" gracePeriod=30 Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.555591 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" podUID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerName="barbican-keystone-listener" containerID="cri-o://5ac62896faa25edf067e9eba6f68e421a1b233d5e85b4ee414fe3961dc9ac78d" gracePeriod=30 Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.565747 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6bdfcdc49-v4hgd" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9696/\": dial tcp 10.217.0.156:9696: connect: connection refused" Nov 28 15:49:35 crc kubenswrapper[4805]: E1128 15:49:35.726748 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:35 crc kubenswrapper[4805]: E1128 15:49:35.727143 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:37.727122132 +0000 UTC m=+1404.776913443 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-config-data" not found Nov 28 15:49:35 crc kubenswrapper[4805]: E1128 15:49:35.726897 4805 secret.go:188] Couldn't get secret openstack/barbican-keystone-listener-config-data: secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:35 crc kubenswrapper[4805]: E1128 15:49:35.727443 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:37.72742963 +0000 UTC m=+1404.777220941 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.953506 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-config-data" (OuterVolumeSpecName: "config-data") pod "e50653a2-4881-42fd-bc8a-35e5f8e2f48d" (UID: "e50653a2-4881-42fd-bc8a-35e5f8e2f48d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:35 crc kubenswrapper[4805]: I1128 15:49:35.978941 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33489fb0-b55a-4544-8a36-c4e0c9ca10f0" (UID: "33489fb0-b55a-4544-8a36-c4e0c9ca10f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.004642 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "e50653a2-4881-42fd-bc8a-35e5f8e2f48d" (UID: "e50653a2-4881-42fd-bc8a-35e5f8e2f48d"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.006234 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "33489fb0-b55a-4544-8a36-c4e0c9ca10f0" (UID: "33489fb0-b55a-4544-8a36-c4e0c9ca10f0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.015453 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "33489fb0-b55a-4544-8a36-c4e0c9ca10f0" (UID: "33489fb0-b55a-4544-8a36-c4e0c9ca10f0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.034571 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-config-data" (OuterVolumeSpecName: "config-data") pod "33489fb0-b55a-4544-8a36-c4e0c9ca10f0" (UID: "33489fb0-b55a-4544-8a36-c4e0c9ca10f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.039853 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.039888 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.039900 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.039910 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.039922 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33489fb0-b55a-4544-8a36-c4e0c9ca10f0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.039934 4805 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.040032 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.040090 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data podName:fb122aae-0e09-46b2-926c-037d25e79477 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:40.04007097 +0000 UTC m=+1407.089862281 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb122aae-0e09-46b2-926c-037d25e79477") : configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.046802 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "e50653a2-4881-42fd-bc8a-35e5f8e2f48d" (UID: "e50653a2-4881-42fd-bc8a-35e5f8e2f48d"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.145096 4805 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50653a2-4881-42fd-bc8a-35e5f8e2f48d-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.191731 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.163:8776/healthcheck\": read tcp 10.217.0.2:40180->10.217.0.163:8776: read: connection reset by peer" Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.546845 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1 is running failed: container process not found" containerID="273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.551546 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1 is running failed: container process not found" containerID="273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.553088 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1 is running failed: container process not found" containerID="273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.553227 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="51493fc6-cc57-4bbf-a6a8-0864160ade3f" containerName="nova-cell1-conductor-conductor" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.608927 4805 generic.go:334] "Generic (PLEG): container finished" podID="ecd6fbaf-5702-415e-8a1d-85d2c4792031" containerID="5bb454b44321bc150d612a41130b3c944d8a6a48515220771451697975c341b2" exitCode=0 Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.621575 4805 generic.go:334] "Generic (PLEG): container finished" podID="51493fc6-cc57-4bbf-a6a8-0864160ade3f" containerID="273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1" exitCode=0 Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.664491 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f is running failed: container process not found" containerID="8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.664643 4805 generic.go:334] "Generic (PLEG): container finished" podID="f912081e-175b-410c-bbfa-daa3a71e2179" containerID="e210a202934599ab4809d38358f193e2eb1eb77d9f8eb89fa27b616eca30451d" exitCode=0 Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.668287 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f is running failed: container process not found" containerID="8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.684005 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f is running failed: container process not found" containerID="8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.684071 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f is running failed: container process not found" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="452c347f-4ee3-46de-ba8f-c83300966f5d" containerName="galera" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.700981 4805 generic.go:334] "Generic (PLEG): container finished" podID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerID="0cc217baf0960656d6ad49571b76347e41bfd409a11a61ef61f68ae6d1b9a1c5" exitCode=0 Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.717047 4805 generic.go:334] "Generic (PLEG): container finished" podID="9d7a9256-533a-4985-9a62-9bbe76af9e54" containerID="b53b84616761467f8383fb98e17f379ec12fbb906d9bd0cffe2c80f86e68acfb" exitCode=0 Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.731999 4805 generic.go:334] "Generic (PLEG): container finished" podID="1f9f5677-6d82-4ee0-bc2e-74c1671be521" containerID="0e6f68b4ef1f21b2f8bb9ebc77214db715f155aaed05e0a61d2a59569c636bbf" exitCode=0 Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.751962 4805 generic.go:334] "Generic (PLEG): container finished" podID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerID="fe684bfa9591b55be6b83497c5f2059493fe4abdfa52867aaec588e8310d8893" exitCode=0 Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.783805 4805 generic.go:334] "Generic (PLEG): container finished" podID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerID="589d702195e92daa9f00af6a62686aa77c68b1a505d4b127244e5ebcb17036cd" exitCode=143 Nov 28 15:49:36 crc kubenswrapper[4805]: E1128 15:49:36.788708 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf912081e_175b_410c_bbfa_daa3a71e2179.slice/crio-e210a202934599ab4809d38358f193e2eb1eb77d9f8eb89fa27b616eca30451d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d7a9256_533a_4985_9a62_9bbe76af9e54.slice/crio-conmon-b53b84616761467f8383fb98e17f379ec12fbb906d9bd0cffe2c80f86e68acfb.scope\": RecentStats: unable to find data in memory cache]" Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.800729 4805 generic.go:334] "Generic (PLEG): container finished" podID="1db4967f-554e-4336-913a-fcebe420cf0e" containerID="a92a82d6fc0b62799e0012cf79d6be83a592db16c726700d8557140daac5fbef" exitCode=0 Nov 28 15:49:36 crc kubenswrapper[4805]: I1128 15:49:36.810714 4805 generic.go:334] "Generic (PLEG): container finished" podID="cf2e5369-a11b-4150-a291-4aeab9724d82" containerID="8af92c9e96307e97e3180f6f4e515c45e8d404b658b8fecda5e22417d08f76db" exitCode=0 Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.014636 4805 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.807s" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.014672 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.014696 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015338 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nv5zc"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015386 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican8232-account-delete-52zhn"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015402 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementbd4b-account-delete-7bztm" event={"ID":"1db4967f-554e-4336-913a-fcebe420cf0e","Type":"ContainerStarted","Data":"570c6cd900b4027822aaea393078090b0aa91ff67f91667fd3dafe88467fcc98"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015423 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015442 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015462 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f8fc6d4f-g6nbp" event={"ID":"33489fb0-b55a-4544-8a36-c4e0c9ca10f0","Type":"ContainerDied","Data":"b5b41c3e7b2867ed3df58372ba9fa27ee74b0e5c928fe954c5aa4b3080b78fb5"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015481 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder72ad-account-delete-wnr85" event={"ID":"f912081e-175b-410c-bbfa-daa3a71e2179","Type":"ContainerStarted","Data":"43ef2e24a7fed0cbd317e912ee73a5e94f395a6852f2a523cc0b0005666c7849"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015496 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015509 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ct546"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015523 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ct546"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.015539 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystonefa91-account-delete-rcm8w"] Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017401 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerName="proxy-httpd" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017644 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerName="proxy-httpd" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017673 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerName="proxy-server" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017683 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerName="proxy-server" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017699 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50653a2-4881-42fd-bc8a-35e5f8e2f48d" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017708 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50653a2-4881-42fd-bc8a-35e5f8e2f48d" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017730 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" containerName="dnsmasq-dns" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017738 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" containerName="dnsmasq-dns" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017755 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerName="ovsdbserver-nb" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017764 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerName="ovsdbserver-nb" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017781 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerName="openstack-network-exporter" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017790 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerName="openstack-network-exporter" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017802 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aff7621-5632-499d-955d-7ea5b9915b54" containerName="openstack-network-exporter" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017809 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aff7621-5632-499d-955d-7ea5b9915b54" containerName="openstack-network-exporter" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017823 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerName="openstack-network-exporter" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017830 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerName="openstack-network-exporter" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017839 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerName="ovsdbserver-sb" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017846 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerName="ovsdbserver-sb" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.017857 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" containerName="init" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.017864 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" containerName="init" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018109 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerName="ovsdbserver-sb" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018136 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerName="openstack-network-exporter" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018145 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" containerName="dnsmasq-dns" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018161 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerName="proxy-httpd" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018175 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" containerName="proxy-server" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018194 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50653a2-4881-42fd-bc8a-35e5f8e2f48d" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018206 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aff7621-5632-499d-955d-7ea5b9915b54" containerName="openstack-network-exporter" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018218 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" containerName="openstack-network-exporter" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018229 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bee11cf6-ec1b-40bb-8030-044d8f06591e" containerName="ovsdbserver-nb" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018880 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-vqdvs"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018899 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5f69cf9fcb-rvfkd"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018923 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-vqdvs"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018938 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp" event={"ID":"b36ae016-8ef7-4c1f-a902-9e6f33d9aaad","Type":"ContainerDied","Data":"af65fb6f9e2b6b005e2729ef460edbd9f67cab838751543c40ada167f3fdaa7d"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018962 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell08782-account-delete-phg59" event={"ID":"ecd6fbaf-5702-415e-8a1d-85d2c4792031","Type":"ContainerStarted","Data":"54d1373449de9e63a701b72d93d3f049df833c6829ad0743cd08f91194871936"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018979 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystonefa91-account-delete-rcm8w"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.018991 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019010 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-vkkgt"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019021 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv5zc" event={"ID":"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5","Type":"ContainerStarted","Data":"8b9784db64b8a1f7b20be90265955171bb18713e7d11151b51e3d2713273a642"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019034 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance73a6-account-delete-qjlt7" event={"ID":"cf2e5369-a11b-4150-a291-4aeab9724d82","Type":"ContainerStarted","Data":"7801d4719043577d7f80f1ec7425157934dda881fde1891589a1b9990a282b68"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019050 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-vkkgt"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019068 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell08782-account-delete-phg59"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019080 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-8782-account-create-update-p6l42"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019091 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vssvf" event={"ID":"89d793e5-c2d1-4630-95ff-615cd30b5d04","Type":"ContainerDied","Data":"2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019106 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7849b77bf-z2qp2" event={"ID":"95167e93-42b8-4f5f-b5a9-587b9b854f12","Type":"ContainerStarted","Data":"fb34bb50b63f1a2069a8960d2e187651a99a18f19f886a0e8973572b03b7950f"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019122 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-8782-account-create-update-p6l42"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019221 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-vf6sz"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019265 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic2b1-account-delete-wfngd" event={"ID":"9d7a9256-533a-4985-9a62-9bbe76af9e54","Type":"ContainerStarted","Data":"47a72f32c1d827f7d0013132d7511e7b4f34f9dadbafbc8bbc216fa1f60da2a1"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019281 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-vf6sz"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019295 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron1f9e-account-delete-vklns" event={"ID":"1f9f5677-6d82-4ee0-bc2e-74c1671be521","Type":"ContainerStarted","Data":"d81354992a8c0a17cb7d14a2b64856356fd98e98df5111476d121afdd30da0a4"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019307 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e50653a2-4881-42fd-bc8a-35e5f8e2f48d","Type":"ContainerDied","Data":"7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019321 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e50653a2-4881-42fd-bc8a-35e5f8e2f48d","Type":"ContainerDied","Data":"4a83f7ce8e72ffffef1ba1d028145e60ff20f88f8ded862d0fdde4857638014f"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019333 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56786c7f4d-vg252" event={"ID":"659e741d-dc3f-4463-b368-075d37a3ef0f","Type":"ContainerStarted","Data":"bba6a6039d9c7d4ae4bc9035b58edee1014582a984d98d9604b06f7f1ef23f2c"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019346 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" event={"ID":"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6","Type":"ContainerDied","Data":"e4e089980980ac67dc67f214af92d43970eb148b6a828b8e14ba26fab9d1df86"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019374 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"452c347f-4ee3-46de-ba8f-c83300966f5d","Type":"ContainerDied","Data":"8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019388 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"452c347f-4ee3-46de-ba8f-c83300966f5d","Type":"ContainerDied","Data":"5c530e830df82888bd8da30d5e625650f7e68c8b030510bac012579aa019d8e4"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019400 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c530e830df82888bd8da30d5e625650f7e68c8b030510bac012579aa019d8e4" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019445 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8232-account-delete-52zhn" event={"ID":"2849f80f-ed31-4c7f-8f65-1132aa35b6e3","Type":"ContainerStarted","Data":"a0811251e175347c191f76c64c954e29e72a8d39c458b0611310ffcaa4e72667"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019458 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell08782-account-delete-phg59" event={"ID":"ecd6fbaf-5702-415e-8a1d-85d2c4792031","Type":"ContainerDied","Data":"5bb454b44321bc150d612a41130b3c944d8a6a48515220771451697975c341b2"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019474 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"51493fc6-cc57-4bbf-a6a8-0864160ade3f","Type":"ContainerDied","Data":"273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019488 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56786c7f4d-vg252" event={"ID":"659e741d-dc3f-4463-b368-075d37a3ef0f","Type":"ContainerStarted","Data":"cbaaf427e82b90388021256c22b24bac4a5e83fbe17c6b89d768e25df50d48c0"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019498 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder72ad-account-delete-wnr85" event={"ID":"f912081e-175b-410c-bbfa-daa3a71e2179","Type":"ContainerDied","Data":"e210a202934599ab4809d38358f193e2eb1eb77d9f8eb89fa27b616eca30451d"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019510 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be5d0310-fb90-4d73-94f3-57b105a2a408","Type":"ContainerDied","Data":"0cc217baf0960656d6ad49571b76347e41bfd409a11a61ef61f68ae6d1b9a1c5"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019522 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic2b1-account-delete-wfngd" event={"ID":"9d7a9256-533a-4985-9a62-9bbe76af9e54","Type":"ContainerDied","Data":"b53b84616761467f8383fb98e17f379ec12fbb906d9bd0cffe2c80f86e68acfb"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019534 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron1f9e-account-delete-vklns" event={"ID":"1f9f5677-6d82-4ee0-bc2e-74c1671be521","Type":"ContainerDied","Data":"0e6f68b4ef1f21b2f8bb9ebc77214db715f155aaed05e0a61d2a59569c636bbf"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019549 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55d54b965d-bfqn6" event={"ID":"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6","Type":"ContainerDied","Data":"fe684bfa9591b55be6b83497c5f2059493fe4abdfa52867aaec588e8310d8893"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019562 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55d54b965d-bfqn6" event={"ID":"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6","Type":"ContainerDied","Data":"1de6d6193c41094be9eb487181477d290298455ffdb3ad015fa40fffe8dd265f"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019572 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1de6d6193c41094be9eb487181477d290298455ffdb3ad015fa40fffe8dd265f" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019583 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" event={"ID":"bbc55e90-2c22-4329-b3cc-375a68e6fe26","Type":"ContainerDied","Data":"589d702195e92daa9f00af6a62686aa77c68b1a505d4b127244e5ebcb17036cd"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019596 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementbd4b-account-delete-7bztm" event={"ID":"1db4967f-554e-4336-913a-fcebe420cf0e","Type":"ContainerDied","Data":"a92a82d6fc0b62799e0012cf79d6be83a592db16c726700d8557140daac5fbef"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019608 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance73a6-account-delete-qjlt7" event={"ID":"cf2e5369-a11b-4150-a291-4aeab9724d82","Type":"ContainerDied","Data":"8af92c9e96307e97e3180f6f4e515c45e8d404b658b8fecda5e22417d08f76db"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.019621 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7849b77bf-z2qp2" event={"ID":"95167e93-42b8-4f5f-b5a9-587b9b854f12","Type":"ContainerStarted","Data":"3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.021287 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="ceilometer-central-agent" containerID="cri-o://0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c" gracePeriod=30 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.021464 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="proxy-httpd" containerID="cri-o://6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8" gracePeriod=30 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.021511 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="sg-core" containerID="cri-o://e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900" gracePeriod=30 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.021543 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="ceilometer-notification-agent" containerID="cri-o://59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a" gracePeriod=30 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.021674 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" containerName="memcached" containerID="cri-o://01f4eeccdd9eae514a9fe127c422ee73a939f8ea3e4c0008f73fe1ea10f2ee26" gracePeriod=30 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.021776 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="aefb91a4-432a-4c5d-8d28-dee413c660c3" containerName="kube-state-metrics" containerID="cri-o://af248f060f9652198f3fb50557ebfd252fb89de3f39e24268855915e1ebf72bd" gracePeriod=30 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.021819 4805 scope.go:117] "RemoveContainer" containerID="a02a9bd5b077fb94a3c91b016fbc51400e43005c07086b7606cdb1ad65200db2" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.021979 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.022341 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-5f69cf9fcb-rvfkd" podUID="614a68ec-3129-413f-abb1-40a73ad9137e" containerName="keystone-api" containerID="cri-o://28e9827ec5a663c00f28c55d8c943235e37131e6dc45f5d050381a2a732dd392" gracePeriod=30 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.046506 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystonefa91-account-delete-rcm8w"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.055730 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-fa91-account-create-update-k9xs4"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.070390 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-fa91-account-create-update-k9xs4"] Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.102547 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.102641 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data podName:5c4171ee-19b0-44c4-8e98-d14e11b9e1a6 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:41.102622816 +0000 UTC m=+1408.152414127 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data") pod "barbican-api-5fc4b9b6bd-9nsgk" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6") : secret "barbican-config-data" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.102720 4805 secret.go:188] Couldn't get secret openstack/barbican-api-config-data: secret "barbican-api-config-data" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.102755 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom podName:5c4171ee-19b0-44c4-8e98-d14e11b9e1a6 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:41.102746259 +0000 UTC m=+1408.152537580 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom") pod "barbican-api-5fc4b9b6bd-9nsgk" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6") : secret "barbican-api-config-data" not found Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.160906 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:51024->10.217.0.200:8775: read: connection reset by peer" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.161035 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:51036->10.217.0.200:8775: read: connection reset by peer" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.204188 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts\") pod \"keystonefa91-account-delete-rcm8w\" (UID: \"aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7\") " pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.204233 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ntr2\" (UniqueName: \"kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2\") pod \"keystonefa91-account-delete-rcm8w\" (UID: \"aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7\") " pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.232859 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07201b49-2b7c-4e44-a6a1-3a2dec9df8bd" path="/var/lib/kubelet/pods/07201b49-2b7c-4e44-a6a1-3a2dec9df8bd/volumes" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.233574 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1232c9a3-93bb-4956-8e76-89829ee8b299" path="/var/lib/kubelet/pods/1232c9a3-93bb-4956-8e76-89829ee8b299/volumes" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.234122 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2efa09bc-8af5-473d-a893-0cdd3936a6ce" path="/var/lib/kubelet/pods/2efa09bc-8af5-473d-a893-0cdd3936a6ce/volumes" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.238637 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67c9ab51-9cfd-4e5f-9df6-1f226ab94173" path="/var/lib/kubelet/pods/67c9ab51-9cfd-4e5f-9df6-1f226ab94173/volumes" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.239144 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac018eb-cb26-4d12-ac87-c200b20a9327" path="/var/lib/kubelet/pods/7ac018eb-cb26-4d12-ac87-c200b20a9327/volumes" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.239644 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4bba869-bfb4-44aa-9510-6dfb82797bf5" path="/var/lib/kubelet/pods/b4bba869-bfb4-44aa-9510-6dfb82797bf5/volumes" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.241337 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc4053dd-426b-473b-9ac2-039a55ab62a7" path="/var/lib/kubelet/pods/bc4053dd-426b-473b-9ac2-039a55ab62a7/volumes" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.248520 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b is running failed: container process not found" containerID="fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.249059 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b is running failed: container process not found" containerID="fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.249540 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b is running failed: container process not found" containerID="fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.249568 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ac086fdf-9a1a-40ab-ab2f-8a33dae28290" containerName="nova-scheduler-scheduler" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.307942 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts\") pod \"keystonefa91-account-delete-rcm8w\" (UID: \"aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7\") " pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.308000 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ntr2\" (UniqueName: \"kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2\") pod \"keystonefa91-account-delete-rcm8w\" (UID: \"aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7\") " pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.308205 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.308296 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts podName:aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:37.808273666 +0000 UTC m=+1404.858065047 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts") pod "keystonefa91-account-delete-rcm8w" (UID: "aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7") : configmap "openstack-scripts" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.314160 4805 projected.go:194] Error preparing data for projected volume kube-api-access-5ntr2 for pod openstack/keystonefa91-account-delete-rcm8w: failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.314244 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2 podName:aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:37.814223649 +0000 UTC m=+1404.864014970 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-5ntr2" (UniqueName: "kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2") pod "keystonefa91-account-delete-rcm8w" (UID: "aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7") : failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.358945 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" containerName="galera" containerID="cri-o://d649033916f10c5d8b5e440164330092d201d1409da92cf8c71f690663343e7c" gracePeriod=30 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.413406 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.419518 4805 scope.go:117] "RemoveContainer" containerID="f341ac5588a30f34bf1c5785f352e9e8bcbb991ecae2e3a87e2814de74b6c3c7" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.512568 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-galera-tls-certs\") pod \"452c347f-4ee3-46de-ba8f-c83300966f5d\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.512994 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-kolla-config\") pod \"452c347f-4ee3-46de-ba8f-c83300966f5d\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.513029 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-combined-ca-bundle\") pod \"452c347f-4ee3-46de-ba8f-c83300966f5d\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.513063 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-default\") pod \"452c347f-4ee3-46de-ba8f-c83300966f5d\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.513150 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smwfr\" (UniqueName: \"kubernetes.io/projected/452c347f-4ee3-46de-ba8f-c83300966f5d-kube-api-access-smwfr\") pod \"452c347f-4ee3-46de-ba8f-c83300966f5d\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.513175 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-operator-scripts\") pod \"452c347f-4ee3-46de-ba8f-c83300966f5d\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.513215 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-generated\") pod \"452c347f-4ee3-46de-ba8f-c83300966f5d\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.513248 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"452c347f-4ee3-46de-ba8f-c83300966f5d\" (UID: \"452c347f-4ee3-46de-ba8f-c83300966f5d\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.513711 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "452c347f-4ee3-46de-ba8f-c83300966f5d" (UID: "452c347f-4ee3-46de-ba8f-c83300966f5d"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.513872 4805 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.514996 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "452c347f-4ee3-46de-ba8f-c83300966f5d" (UID: "452c347f-4ee3-46de-ba8f-c83300966f5d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.515236 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "452c347f-4ee3-46de-ba8f-c83300966f5d" (UID: "452c347f-4ee3-46de-ba8f-c83300966f5d"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.515434 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "452c347f-4ee3-46de-ba8f-c83300966f5d" (UID: "452c347f-4ee3-46de-ba8f-c83300966f5d"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.534754 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/452c347f-4ee3-46de-ba8f-c83300966f5d-kube-api-access-smwfr" (OuterVolumeSpecName: "kube-api-access-smwfr") pod "452c347f-4ee3-46de-ba8f-c83300966f5d" (UID: "452c347f-4ee3-46de-ba8f-c83300966f5d"). InnerVolumeSpecName "kube-api-access-smwfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.554920 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "452c347f-4ee3-46de-ba8f-c83300966f5d" (UID: "452c347f-4ee3-46de-ba8f-c83300966f5d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.576844 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "452c347f-4ee3-46de-ba8f-c83300966f5d" (UID: "452c347f-4ee3-46de-ba8f-c83300966f5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.615763 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.615842 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.615853 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smwfr\" (UniqueName: \"kubernetes.io/projected/452c347f-4ee3-46de-ba8f-c83300966f5d-kube-api-access-smwfr\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.615865 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/452c347f-4ee3-46de-ba8f-c83300966f5d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.615875 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/452c347f-4ee3-46de-ba8f-c83300966f5d-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.615896 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.663956 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.684680 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "452c347f-4ee3-46de-ba8f-c83300966f5d" (UID: "452c347f-4ee3-46de-ba8f-c83300966f5d"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.718100 4805 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/452c347f-4ee3-46de-ba8f-c83300966f5d-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.718132 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.726124 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.153:9311/healthcheck\": read tcp 10.217.0.2:53190->10.217.0.153:9311: read: connection reset by peer" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.726198 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.153:9311/healthcheck\": read tcp 10.217.0.2:53186->10.217.0.153:9311: read: connection reset by peer" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.739682 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.761021 4805 scope.go:117] "RemoveContainer" containerID="051cad31407ced2cb3d554ef9c21f2839c03af141e5fd2cda0c2a783a4cd32a3" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.761664 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.815173 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.820974 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts\") pod \"keystonefa91-account-delete-rcm8w\" (UID: \"aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7\") " pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.821019 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ntr2\" (UniqueName: \"kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2\") pod \"keystonefa91-account-delete-rcm8w\" (UID: \"aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7\") " pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.821195 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.821241 4805 secret.go:188] Couldn't get secret openstack/barbican-keystone-listener-config-data: secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.821267 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts podName:aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:38.821251281 +0000 UTC m=+1405.871042592 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts") pod "keystonefa91-account-delete-rcm8w" (UID: "aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7") : configmap "openstack-scripts" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.821284 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:41.821277461 +0000 UTC m=+1408.871068772 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.821381 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.821407 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:41.821400365 +0000 UTC m=+1408.871191676 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-config-data" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.827102 4805 projected.go:194] Error preparing data for projected volume kube-api-access-5ntr2 for pod openstack/keystonefa91-account-delete-rcm8w: failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.827163 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2 podName:aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:38.827145732 +0000 UTC m=+1405.876937043 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-5ntr2" (UniqueName: "kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2") pod "keystonefa91-account-delete-rcm8w" (UID: "aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7") : failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 15:49:37 crc kubenswrapper[4805]: E1128 15:49:37.830021 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-5ntr2 operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystonefa91-account-delete-rcm8w" podUID="aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.841162 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.853017 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-7jmkp"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.859042 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.859583 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6f8fc6d4f-g6nbp"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.863544 4805 generic.go:334] "Generic (PLEG): container finished" podID="bd26b960-0994-4bb3-aef3-035519e32420" containerID="e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900" exitCode=2 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.863620 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerDied","Data":"e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.866597 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-6f8fc6d4f-g6nbp"] Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.873717 4805 generic.go:334] "Generic (PLEG): container finished" podID="ac086fdf-9a1a-40ab-ab2f-8a33dae28290" containerID="fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b" exitCode=0 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.873810 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ac086fdf-9a1a-40ab-ab2f-8a33dae28290","Type":"ContainerDied","Data":"fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.873839 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ac086fdf-9a1a-40ab-ab2f-8a33dae28290","Type":"ContainerDied","Data":"b5e2786e52c751fdc5d474871b9e78d5a30793ccb2a30907a1d2491c251f095d"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.873854 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5e2786e52c751fdc5d474871b9e78d5a30793ccb2a30907a1d2491c251f095d" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.883593 4805 generic.go:334] "Generic (PLEG): container finished" podID="d7348256-32de-4696-9637-96d7323831ed" containerID="5b7e15d112969a7b953c22fe09274e56dd365f74ce49e846d9443a14feb3ce7a" exitCode=0 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.883662 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7348256-32de-4696-9637-96d7323831ed","Type":"ContainerDied","Data":"5b7e15d112969a7b953c22fe09274e56dd365f74ce49e846d9443a14feb3ce7a"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.884158 4805 scope.go:117] "RemoveContainer" containerID="f512ffb73a4fae0efab8bd72a241efe3ae50f7565449d1f938e5a9c88c925c05" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.903818 4805 generic.go:334] "Generic (PLEG): container finished" podID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerID="02772248390f9aa46863e063f9eac9961344ec1704c67c8cfde40db99e473634" exitCode=0 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.903882 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" event={"ID":"a363e2d1-0a91-4578-b9e4-b736b7931b03","Type":"ContainerDied","Data":"02772248390f9aa46863e063f9eac9961344ec1704c67c8cfde40db99e473634"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.903905 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" event={"ID":"a363e2d1-0a91-4578-b9e4-b736b7931b03","Type":"ContainerDied","Data":"c1900187afec2dc4e77cef6a31189c5d8ad610ee80412555fe4dfc6358cea286"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.903916 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1900187afec2dc4e77cef6a31189c5d8ad610ee80412555fe4dfc6358cea286" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.913631 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.917786 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be5d0310-fb90-4d73-94f3-57b105a2a408","Type":"ContainerDied","Data":"8f5e0f1d7c83ce9911dbbe2425dbef50ab47911c099ee055d4c347c824de0de7"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.917846 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926078 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-internal-tls-certs\") pod \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926144 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be5d0310-fb90-4d73-94f3-57b105a2a408-etc-machine-id\") pod \"be5d0310-fb90-4d73-94f3-57b105a2a408\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926206 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-combined-ca-bundle\") pod \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926225 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-combined-ca-bundle\") pod \"a363e2d1-0a91-4578-b9e4-b736b7931b03\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926287 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-public-tls-certs\") pod \"be5d0310-fb90-4d73-94f3-57b105a2a408\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926364 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrqwf\" (UniqueName: \"kubernetes.io/projected/6ef711de-53e6-4705-8d2f-469b6dc2d4de-kube-api-access-qrqwf\") pod \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926446 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf67w\" (UniqueName: \"kubernetes.io/projected/be5d0310-fb90-4d73-94f3-57b105a2a408-kube-api-access-tf67w\") pod \"be5d0310-fb90-4d73-94f3-57b105a2a408\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926466 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data-custom\") pod \"a363e2d1-0a91-4578-b9e4-b736b7931b03\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926486 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-config-data\") pod \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926557 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st67h\" (UniqueName: \"kubernetes.io/projected/a363e2d1-0a91-4578-b9e4-b736b7931b03-kube-api-access-st67h\") pod \"a363e2d1-0a91-4578-b9e4-b736b7931b03\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926612 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-combined-ca-bundle\") pod \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926632 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-internal-tls-certs\") pod \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926674 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5d0310-fb90-4d73-94f3-57b105a2a408-logs\") pod \"be5d0310-fb90-4d73-94f3-57b105a2a408\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926741 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-combined-ca-bundle\") pod \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926765 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-public-tls-certs\") pod \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926821 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ef711de-53e6-4705-8d2f-469b6dc2d4de-logs\") pod \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926857 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-config-data\") pod \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\" (UID: \"6ef711de-53e6-4705-8d2f-469b6dc2d4de\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926899 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-internal-tls-certs\") pod \"be5d0310-fb90-4d73-94f3-57b105a2a408\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926922 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a363e2d1-0a91-4578-b9e4-b736b7931b03-logs\") pod \"a363e2d1-0a91-4578-b9e4-b736b7931b03\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926945 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data\") pod \"be5d0310-fb90-4d73-94f3-57b105a2a408\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.926989 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hldcl\" (UniqueName: \"kubernetes.io/projected/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-kube-api-access-hldcl\") pod \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.927022 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-logs\") pod \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.927086 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-public-tls-certs\") pod \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.927143 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-combined-ca-bundle\") pod \"be5d0310-fb90-4d73-94f3-57b105a2a408\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.927171 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-scripts\") pod \"be5d0310-fb90-4d73-94f3-57b105a2a408\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.927218 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xghst\" (UniqueName: \"kubernetes.io/projected/51493fc6-cc57-4bbf-a6a8-0864160ade3f-kube-api-access-xghst\") pod \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\" (UID: \"51493fc6-cc57-4bbf-a6a8-0864160ade3f\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.927263 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-scripts\") pod \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.927308 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data-custom\") pod \"be5d0310-fb90-4d73-94f3-57b105a2a408\" (UID: \"be5d0310-fb90-4d73-94f3-57b105a2a408\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.927329 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-config-data\") pod \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\" (UID: \"0e72cc62-edd5-4d0c-89a8-099d6b07e1e6\") " Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.928820 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ef711de-53e6-4705-8d2f-469b6dc2d4de-logs" (OuterVolumeSpecName: "logs") pod "6ef711de-53e6-4705-8d2f-469b6dc2d4de" (UID: "6ef711de-53e6-4705-8d2f-469b6dc2d4de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.929505 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ef711de-53e6-4705-8d2f-469b6dc2d4de-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.929729 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be5d0310-fb90-4d73-94f3-57b105a2a408-logs" (OuterVolumeSpecName: "logs") pod "be5d0310-fb90-4d73-94f3-57b105a2a408" (UID: "be5d0310-fb90-4d73-94f3-57b105a2a408"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.930385 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be5d0310-fb90-4d73-94f3-57b105a2a408-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "be5d0310-fb90-4d73-94f3-57b105a2a408" (UID: "be5d0310-fb90-4d73-94f3-57b105a2a408"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.939341 4805 scope.go:117] "RemoveContainer" containerID="0c5162ba13d3afef2eddf48abd4934d9802217a483faa8ade07f7b373cd99503" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.946911 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a363e2d1-0a91-4578-b9e4-b736b7931b03-logs" (OuterVolumeSpecName: "logs") pod "a363e2d1-0a91-4578-b9e4-b736b7931b03" (UID: "a363e2d1-0a91-4578-b9e4-b736b7931b03"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.951813 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-logs" (OuterVolumeSpecName: "logs") pod "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" (UID: "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.960512 4805 generic.go:334] "Generic (PLEG): container finished" podID="aefb91a4-432a-4c5d-8d28-dee413c660c3" containerID="af248f060f9652198f3fb50557ebfd252fb89de3f39e24268855915e1ebf72bd" exitCode=2 Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.960569 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aefb91a4-432a-4c5d-8d28-dee413c660c3","Type":"ContainerDied","Data":"af248f060f9652198f3fb50557ebfd252fb89de3f39e24268855915e1ebf72bd"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.960594 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aefb91a4-432a-4c5d-8d28-dee413c660c3","Type":"ContainerDied","Data":"be518e9ace91a19bf38638deafca21d394a18966a9f0499d4f0f3adb3830e2ad"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.960606 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be518e9ace91a19bf38638deafca21d394a18966a9f0499d4f0f3adb3830e2ad" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.960979 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-kube-api-access-hldcl" (OuterVolumeSpecName: "kube-api-access-hldcl") pod "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" (UID: "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6"). InnerVolumeSpecName "kube-api-access-hldcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.964279 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.977448 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.978565 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"51493fc6-cc57-4bbf-a6a8-0864160ade3f","Type":"ContainerDied","Data":"f02f5af79881f215334652ad50e645be8b7f4449bd67bc35806c5d03b49a41a0"} Nov 28 15:49:37 crc kubenswrapper[4805]: I1128 15:49:37.982310 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.990942 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.992554 4805 generic.go:334] "Generic (PLEG): container finished" podID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerID="dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649" exitCode=0 Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.992606 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ef711de-53e6-4705-8d2f-469b6dc2d4de","Type":"ContainerDied","Data":"dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649"} Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.992629 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ef711de-53e6-4705-8d2f-469b6dc2d4de","Type":"ContainerDied","Data":"594f4fffd996d60bb888468f4ca60b1e16c97ad5601423a91ec7c755ceda4f9f"} Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.992670 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.996282 4805 generic.go:334] "Generic (PLEG): container finished" podID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerID="31105a43fdab9daa89f1ca09a13406cb4b8904817a89036a2573227173287ae5" exitCode=0 Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.996324 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c021413-239d-4ba0-979d-1c4b25d5093c","Type":"ContainerDied","Data":"31105a43fdab9daa89f1ca09a13406cb4b8904817a89036a2573227173287ae5"} Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.998088 4805 generic.go:334] "Generic (PLEG): container finished" podID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerID="1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de" exitCode=0 Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.998123 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-97ccb4d45-ddn2r" event={"ID":"abfd499c-e83b-4616-a80c-29a7e8e750e2","Type":"ContainerDied","Data":"1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de"} Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.998138 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-97ccb4d45-ddn2r" event={"ID":"abfd499c-e83b-4616-a80c-29a7e8e750e2","Type":"ContainerDied","Data":"0f69434a9db9bceb12c66989fba99a7fce7624aaa43608837cd7f1d51a750edb"} Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:37.998192 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-97ccb4d45-ddn2r" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.000118 4805 generic.go:334] "Generic (PLEG): container finished" podID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerID="868c845c22044f953c99bf70b59a10a7cf406a71b3f9ce6c6262250153d75b30" exitCode=0 Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.000155 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"19ef9cc1-28ab-4014-897c-9679f3d36443","Type":"ContainerDied","Data":"868c845c22044f953c99bf70b59a10a7cf406a71b3f9ce6c6262250153d75b30"} Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.001841 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-scripts" (OuterVolumeSpecName: "scripts") pod "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" (UID: "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.003574 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-scripts" (OuterVolumeSpecName: "scripts") pod "be5d0310-fb90-4d73-94f3-57b105a2a408" (UID: "be5d0310-fb90-4d73-94f3-57b105a2a408"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.004240 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a363e2d1-0a91-4578-b9e4-b736b7931b03-kube-api-access-st67h" (OuterVolumeSpecName: "kube-api-access-st67h") pod "a363e2d1-0a91-4578-b9e4-b736b7931b03" (UID: "a363e2d1-0a91-4578-b9e4-b736b7931b03"). InnerVolumeSpecName "kube-api-access-st67h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.004657 4805 scope.go:117] "RemoveContainer" containerID="7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.006929 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "be5d0310-fb90-4d73-94f3-57b105a2a408" (UID: "be5d0310-fb90-4d73-94f3-57b105a2a408"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.010736 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.020364 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be5d0310-fb90-4d73-94f3-57b105a2a408-kube-api-access-tf67w" (OuterVolumeSpecName: "kube-api-access-tf67w") pod "be5d0310-fb90-4d73-94f3-57b105a2a408" (UID: "be5d0310-fb90-4d73-94f3-57b105a2a408"). InnerVolumeSpecName "kube-api-access-tf67w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.022797 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55d54b965d-bfqn6" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.025499 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ef711de-53e6-4705-8d2f-469b6dc2d4de-kube-api-access-qrqwf" (OuterVolumeSpecName: "kube-api-access-qrqwf") pod "6ef711de-53e6-4705-8d2f-469b6dc2d4de" (UID: "6ef711de-53e6-4705-8d2f-469b6dc2d4de"). InnerVolumeSpecName "kube-api-access-qrqwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.027018 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a363e2d1-0a91-4578-b9e4-b736b7931b03" (UID: "a363e2d1-0a91-4578-b9e4-b736b7931b03"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.033705 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data\") pod \"a363e2d1-0a91-4578-b9e4-b736b7931b03\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.034034 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data-custom\") pod \"a363e2d1-0a91-4578-b9e4-b736b7931b03\" (UID: \"a363e2d1-0a91-4578-b9e4-b736b7931b03\") " Nov 28 15:49:38 crc kubenswrapper[4805]: W1128 15:49:38.035039 4805 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/a363e2d1-0a91-4578-b9e4-b736b7931b03/volumes/kubernetes.io~secret/config-data-custom Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.035065 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a363e2d1-0a91-4578-b9e4-b736b7931b03" (UID: "a363e2d1-0a91-4578-b9e4-b736b7931b03"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.041816 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51493fc6-cc57-4bbf-a6a8-0864160ade3f-kube-api-access-xghst" (OuterVolumeSpecName: "kube-api-access-xghst") pod "51493fc6-cc57-4bbf-a6a8-0864160ade3f" (UID: "51493fc6-cc57-4bbf-a6a8-0864160ade3f"). InnerVolumeSpecName "kube-api-access-xghst". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044658 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a363e2d1-0a91-4578-b9e4-b736b7931b03-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044692 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hldcl\" (UniqueName: \"kubernetes.io/projected/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-kube-api-access-hldcl\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044702 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044712 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044721 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xghst\" (UniqueName: \"kubernetes.io/projected/51493fc6-cc57-4bbf-a6a8-0864160ade3f-kube-api-access-xghst\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044732 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044742 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044753 4805 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be5d0310-fb90-4d73-94f3-57b105a2a408-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044762 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrqwf\" (UniqueName: \"kubernetes.io/projected/6ef711de-53e6-4705-8d2f-469b6dc2d4de-kube-api-access-qrqwf\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044771 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf67w\" (UniqueName: \"kubernetes.io/projected/be5d0310-fb90-4d73-94f3-57b105a2a408-kube-api-access-tf67w\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044779 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044788 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st67h\" (UniqueName: \"kubernetes.io/projected/a363e2d1-0a91-4578-b9e4-b736b7931b03-kube-api-access-st67h\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.044795 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5d0310-fb90-4d73-94f3-57b105a2a408-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.087750 4805 scope.go:117] "RemoveContainer" containerID="7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73" Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.088342 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73\": container with ID starting with 7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73 not found: ID does not exist" containerID="7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.088435 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73"} err="failed to get container status \"7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73\": rpc error: code = NotFound desc = could not find container \"7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73\": container with ID starting with 7bedfed8d1ab5ae73a68ffb0f3e79404bacc926c66c2efbe013a78d99a5d2c73 not found: ID does not exist" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.088456 4805 scope.go:117] "RemoveContainer" containerID="0cc217baf0960656d6ad49571b76347e41bfd409a11a61ef61f68ae6d1b9a1c5" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.102705 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.116842 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.146601 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-combined-ca-bundle\") pod \"abfd499c-e83b-4616-a80c-29a7e8e750e2\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.146665 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-combined-ca-bundle\") pod \"aefb91a4-432a-4c5d-8d28-dee413c660c3\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.146718 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data\") pod \"abfd499c-e83b-4616-a80c-29a7e8e750e2\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.146774 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxdxp\" (UniqueName: \"kubernetes.io/projected/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-api-access-kxdxp\") pod \"aefb91a4-432a-4c5d-8d28-dee413c660c3\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.146825 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-config\") pod \"aefb91a4-432a-4c5d-8d28-dee413c660c3\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.146904 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abfd499c-e83b-4616-a80c-29a7e8e750e2-logs\") pod \"abfd499c-e83b-4616-a80c-29a7e8e750e2\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.146994 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-certs\") pod \"aefb91a4-432a-4c5d-8d28-dee413c660c3\" (UID: \"aefb91a4-432a-4c5d-8d28-dee413c660c3\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.147561 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-combined-ca-bundle\") pod \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.147639 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ftr4\" (UniqueName: \"kubernetes.io/projected/abfd499c-e83b-4616-a80c-29a7e8e750e2-kube-api-access-2ftr4\") pod \"abfd499c-e83b-4616-a80c-29a7e8e750e2\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.147681 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data-custom\") pod \"abfd499c-e83b-4616-a80c-29a7e8e750e2\" (UID: \"abfd499c-e83b-4616-a80c-29a7e8e750e2\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.147716 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl9np\" (UniqueName: \"kubernetes.io/projected/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-kube-api-access-dl9np\") pod \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.147760 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abfd499c-e83b-4616-a80c-29a7e8e750e2-logs" (OuterVolumeSpecName: "logs") pod "abfd499c-e83b-4616-a80c-29a7e8e750e2" (UID: "abfd499c-e83b-4616-a80c-29a7e8e750e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.147812 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-config-data\") pod \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\" (UID: \"ac086fdf-9a1a-40ab-ab2f-8a33dae28290\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.149035 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abfd499c-e83b-4616-a80c-29a7e8e750e2-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.198078 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-api-access-kxdxp" (OuterVolumeSpecName: "kube-api-access-kxdxp") pod "aefb91a4-432a-4c5d-8d28-dee413c660c3" (UID: "aefb91a4-432a-4c5d-8d28-dee413c660c3"). InnerVolumeSpecName "kube-api-access-kxdxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.199766 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "abfd499c-e83b-4616-a80c-29a7e8e750e2" (UID: "abfd499c-e83b-4616-a80c-29a7e8e750e2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.199912 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abfd499c-e83b-4616-a80c-29a7e8e750e2-kube-api-access-2ftr4" (OuterVolumeSpecName: "kube-api-access-2ftr4") pod "abfd499c-e83b-4616-a80c-29a7e8e750e2" (UID: "abfd499c-e83b-4616-a80c-29a7e8e750e2"). InnerVolumeSpecName "kube-api-access-2ftr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.200235 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-kube-api-access-dl9np" (OuterVolumeSpecName: "kube-api-access-dl9np") pod "ac086fdf-9a1a-40ab-ab2f-8a33dae28290" (UID: "ac086fdf-9a1a-40ab-ab2f-8a33dae28290"). InnerVolumeSpecName "kube-api-access-dl9np". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.223146 4805 scope.go:117] "RemoveContainer" containerID="16befe4e2598415b16fc7b244b3cb1d7660ff6773e1cc20677bf867cfd4f1442" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.249792 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-config-data" (OuterVolumeSpecName: "config-data") pod "6ef711de-53e6-4705-8d2f-469b6dc2d4de" (UID: "6ef711de-53e6-4705-8d2f-469b6dc2d4de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.251110 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl9np\" (UniqueName: \"kubernetes.io/projected/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-kube-api-access-dl9np\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.251131 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.251140 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxdxp\" (UniqueName: \"kubernetes.io/projected/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-api-access-kxdxp\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.251149 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ftr4\" (UniqueName: \"kubernetes.io/projected/abfd499c-e83b-4616-a80c-29a7e8e750e2-kube-api-access-2ftr4\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.251157 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.265671 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-config-data" (OuterVolumeSpecName: "config-data") pod "51493fc6-cc57-4bbf-a6a8-0864160ade3f" (UID: "51493fc6-cc57-4bbf-a6a8-0864160ade3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.276119 4805 scope.go:117] "RemoveContainer" containerID="273867062f875771c4f8ebfc8ae8cc4bb2e098ed1da004cad4f280bf7e6140c1" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.306128 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" (UID: "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.316735 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data" (OuterVolumeSpecName: "config-data") pod "abfd499c-e83b-4616-a80c-29a7e8e750e2" (UID: "abfd499c-e83b-4616-a80c-29a7e8e750e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.326595 4805 scope.go:117] "RemoveContainer" containerID="dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.340737 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a363e2d1-0a91-4578-b9e4-b736b7931b03" (UID: "a363e2d1-0a91-4578-b9e4-b736b7931b03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.341024 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-config-data" (OuterVolumeSpecName: "config-data") pod "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" (UID: "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.343531 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be5d0310-fb90-4d73-94f3-57b105a2a408" (UID: "be5d0310-fb90-4d73-94f3-57b105a2a408"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.349605 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51493fc6-cc57-4bbf-a6a8-0864160ade3f" (UID: "51493fc6-cc57-4bbf-a6a8-0864160ade3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.354120 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.354151 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.354165 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.354177 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.354187 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51493fc6-cc57-4bbf-a6a8-0864160ade3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.354198 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.354210 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.357205 4805 scope.go:117] "RemoveContainer" containerID="58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.380988 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-config-data" (OuterVolumeSpecName: "config-data") pod "ac086fdf-9a1a-40ab-ab2f-8a33dae28290" (UID: "ac086fdf-9a1a-40ab-ab2f-8a33dae28290"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.384609 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aefb91a4-432a-4c5d-8d28-dee413c660c3" (UID: "aefb91a4-432a-4c5d-8d28-dee413c660c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.403543 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data" (OuterVolumeSpecName: "config-data") pod "a363e2d1-0a91-4578-b9e4-b736b7931b03" (UID: "a363e2d1-0a91-4578-b9e4-b736b7931b03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.412403 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ef711de-53e6-4705-8d2f-469b6dc2d4de" (UID: "6ef711de-53e6-4705-8d2f-469b6dc2d4de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.423958 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be5d0310-fb90-4d73-94f3-57b105a2a408" (UID: "be5d0310-fb90-4d73-94f3-57b105a2a408"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.433146 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "be5d0310-fb90-4d73-94f3-57b105a2a408" (UID: "be5d0310-fb90-4d73-94f3-57b105a2a408"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.452990 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data" (OuterVolumeSpecName: "config-data") pod "be5d0310-fb90-4d73-94f3-57b105a2a408" (UID: "be5d0310-fb90-4d73-94f3-57b105a2a408"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.454071 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6ef711de-53e6-4705-8d2f-469b6dc2d4de" (UID: "6ef711de-53e6-4705-8d2f-469b6dc2d4de"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.455904 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.455931 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.456030 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.456039 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a363e2d1-0a91-4578-b9e4-b736b7931b03-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.456048 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be5d0310-fb90-4d73-94f3-57b105a2a408-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.456057 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.456065 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.456073 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.469707 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6ef711de-53e6-4705-8d2f-469b6dc2d4de" (UID: "6ef711de-53e6-4705-8d2f-469b6dc2d4de"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.495710 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "aefb91a4-432a-4c5d-8d28-dee413c660c3" (UID: "aefb91a4-432a-4c5d-8d28-dee413c660c3"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.551550 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac086fdf-9a1a-40ab-ab2f-8a33dae28290" (UID: "ac086fdf-9a1a-40ab-ab2f-8a33dae28290"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.558016 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac086fdf-9a1a-40ab-ab2f-8a33dae28290-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.558055 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ef711de-53e6-4705-8d2f-469b6dc2d4de-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.558065 4805 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.643992 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abfd499c-e83b-4616-a80c-29a7e8e750e2" (UID: "abfd499c-e83b-4616-a80c-29a7e8e750e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.659994 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd499c-e83b-4616-a80c-29a7e8e750e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.756192 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "aefb91a4-432a-4c5d-8d28-dee413c660c3" (UID: "aefb91a4-432a-4c5d-8d28-dee413c660c3"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.762036 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" (UID: "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.772729 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.772997 4805 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/aefb91a4-432a-4c5d-8d28-dee413c660c3-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.772820 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.773200 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data podName:4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a nodeName:}" failed. No retries permitted until 2025-11-28 15:49:46.773173196 +0000 UTC m=+1413.822964507 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data") pod "rabbitmq-server-0" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a") : configmap "rabbitmq-config-data" not found Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.773578 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.787500 4805 scope.go:117] "RemoveContainer" containerID="dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649" Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.788281 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649\": container with ID starting with dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649 not found: ID does not exist" containerID="dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.788339 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649"} err="failed to get container status \"dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649\": rpc error: code = NotFound desc = could not find container \"dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649\": container with ID starting with dddb125a0cb07344526453b3646c0a664d2345b78a367ac93191e8b1c894d649 not found: ID does not exist" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.788418 4805 scope.go:117] "RemoveContainer" containerID="58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194" Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.788903 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194\": container with ID starting with 58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194 not found: ID does not exist" containerID="58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.788938 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194"} err="failed to get container status \"58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194\": rpc error: code = NotFound desc = could not find container \"58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194\": container with ID starting with 58c1d2b34cb1f9f949c26633736fc65459896e87bb2392beb319d9699f123194 not found: ID does not exist" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.788957 4805 scope.go:117] "RemoveContainer" containerID="1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.793825 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.797408 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.814823 4805 scope.go:117] "RemoveContainer" containerID="c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.818794 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.819519 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.827183 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.828800 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.833779 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.861299 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" (UID: "0e72cc62-edd5-4d0c-89a8-099d6b07e1e6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.863962 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.873349 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874024 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-logs\") pod \"19ef9cc1-28ab-4014-897c-9679f3d36443\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874065 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-config-data\") pod \"19ef9cc1-28ab-4014-897c-9679f3d36443\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874088 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x45zf\" (UniqueName: \"kubernetes.io/projected/d7348256-32de-4696-9637-96d7323831ed-kube-api-access-x45zf\") pod \"d7348256-32de-4696-9637-96d7323831ed\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874106 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf2e5369-a11b-4150-a291-4aeab9724d82-operator-scripts\") pod \"cf2e5369-a11b-4150-a291-4aeab9724d82\" (UID: \"cf2e5369-a11b-4150-a291-4aeab9724d82\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874130 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-internal-tls-certs\") pod \"19ef9cc1-28ab-4014-897c-9679f3d36443\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874153 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-config-data\") pod \"d7348256-32de-4696-9637-96d7323831ed\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874178 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2s4b\" (UniqueName: \"kubernetes.io/projected/ecd6fbaf-5702-415e-8a1d-85d2c4792031-kube-api-access-w2s4b\") pod \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\" (UID: \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874213 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-nova-metadata-tls-certs\") pod \"d7348256-32de-4696-9637-96d7323831ed\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874232 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7348256-32de-4696-9637-96d7323831ed-logs\") pod \"d7348256-32de-4696-9637-96d7323831ed\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874298 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmfd7\" (UniqueName: \"kubernetes.io/projected/1db4967f-554e-4336-913a-fcebe420cf0e-kube-api-access-qmfd7\") pod \"1db4967f-554e-4336-913a-fcebe420cf0e\" (UID: \"1db4967f-554e-4336-913a-fcebe420cf0e\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874900 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-httpd-run\") pod \"19ef9cc1-28ab-4014-897c-9679f3d36443\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874954 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-combined-ca-bundle\") pod \"19ef9cc1-28ab-4014-897c-9679f3d36443\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.874973 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx5gn\" (UniqueName: \"kubernetes.io/projected/19ef9cc1-28ab-4014-897c-9679f3d36443-kube-api-access-vx5gn\") pod \"19ef9cc1-28ab-4014-897c-9679f3d36443\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875025 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-scripts\") pod \"19ef9cc1-28ab-4014-897c-9679f3d36443\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875046 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-combined-ca-bundle\") pod \"d7348256-32de-4696-9637-96d7323831ed\" (UID: \"d7348256-32de-4696-9637-96d7323831ed\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875043 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-logs" (OuterVolumeSpecName: "logs") pod "19ef9cc1-28ab-4014-897c-9679f3d36443" (UID: "19ef9cc1-28ab-4014-897c-9679f3d36443"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875072 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9mb4\" (UniqueName: \"kubernetes.io/projected/cf2e5369-a11b-4150-a291-4aeab9724d82-kube-api-access-r9mb4\") pod \"cf2e5369-a11b-4150-a291-4aeab9724d82\" (UID: \"cf2e5369-a11b-4150-a291-4aeab9724d82\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875106 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db4967f-554e-4336-913a-fcebe420cf0e-operator-scripts\") pod \"1db4967f-554e-4336-913a-fcebe420cf0e\" (UID: \"1db4967f-554e-4336-913a-fcebe420cf0e\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875148 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd6fbaf-5702-415e-8a1d-85d2c4792031-operator-scripts\") pod \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\" (UID: \"ecd6fbaf-5702-415e-8a1d-85d2c4792031\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875182 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"19ef9cc1-28ab-4014-897c-9679f3d36443\" (UID: \"19ef9cc1-28ab-4014-897c-9679f3d36443\") " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875485 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts\") pod \"keystonefa91-account-delete-rcm8w\" (UID: \"aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7\") " pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875513 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ntr2\" (UniqueName: \"kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2\") pod \"keystonefa91-account-delete-rcm8w\" (UID: \"aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7\") " pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875619 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.875631 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.878226 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf2e5369-a11b-4150-a291-4aeab9724d82-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf2e5369-a11b-4150-a291-4aeab9724d82" (UID: "cf2e5369-a11b-4150-a291-4aeab9724d82"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.882716 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7348256-32de-4696-9637-96d7323831ed-kube-api-access-x45zf" (OuterVolumeSpecName: "kube-api-access-x45zf") pod "d7348256-32de-4696-9637-96d7323831ed" (UID: "d7348256-32de-4696-9637-96d7323831ed"). InnerVolumeSpecName "kube-api-access-x45zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.883839 4805 projected.go:194] Error preparing data for projected volume kube-api-access-5ntr2 for pod openstack/keystonefa91-account-delete-rcm8w: failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.883912 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2 podName:aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:40.883890879 +0000 UTC m=+1407.933682260 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-5ntr2" (UniqueName: "kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2") pod "keystonefa91-account-delete-rcm8w" (UID: "aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7") : failed to fetch token: serviceaccounts "galera-openstack" not found Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.884136 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd6fbaf-5702-415e-8a1d-85d2c4792031-kube-api-access-w2s4b" (OuterVolumeSpecName: "kube-api-access-w2s4b") pod "ecd6fbaf-5702-415e-8a1d-85d2c4792031" (UID: "ecd6fbaf-5702-415e-8a1d-85d2c4792031"). InnerVolumeSpecName "kube-api-access-w2s4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.885458 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecd6fbaf-5702-415e-8a1d-85d2c4792031-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ecd6fbaf-5702-415e-8a1d-85d2c4792031" (UID: "ecd6fbaf-5702-415e-8a1d-85d2c4792031"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.889894 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-scripts" (OuterVolumeSpecName: "scripts") pod "19ef9cc1-28ab-4014-897c-9679f3d36443" (UID: "19ef9cc1-28ab-4014-897c-9679f3d36443"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.890383 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf2e5369-a11b-4150-a291-4aeab9724d82-kube-api-access-r9mb4" (OuterVolumeSpecName: "kube-api-access-r9mb4") pod "cf2e5369-a11b-4150-a291-4aeab9724d82" (UID: "cf2e5369-a11b-4150-a291-4aeab9724d82"). InnerVolumeSpecName "kube-api-access-r9mb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.890791 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1db4967f-554e-4336-913a-fcebe420cf0e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1db4967f-554e-4336-913a-fcebe420cf0e" (UID: "1db4967f-554e-4336-913a-fcebe420cf0e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.891082 4805 scope.go:117] "RemoveContainer" containerID="1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.891152 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "19ef9cc1-28ab-4014-897c-9679f3d36443" (UID: "19ef9cc1-28ab-4014-897c-9679f3d36443"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.891295 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.891340 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7348256-32de-4696-9637-96d7323831ed-logs" (OuterVolumeSpecName: "logs") pod "d7348256-32de-4696-9637-96d7323831ed" (UID: "d7348256-32de-4696-9637-96d7323831ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.891393 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts podName:aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:40.891370014 +0000 UTC m=+1407.941161325 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts") pod "keystonefa91-account-delete-rcm8w" (UID: "aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7") : configmap "openstack-scripts" not found Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.893014 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de\": container with ID starting with 1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de not found: ID does not exist" containerID="1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.893052 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de"} err="failed to get container status \"1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de\": rpc error: code = NotFound desc = could not find container \"1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de\": container with ID starting with 1c86736c01b10789d8cf00ca401a9ceed5b593f6171b5b26feb44a98570f20de not found: ID does not exist" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.893074 4805 scope.go:117] "RemoveContainer" containerID="c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5" Nov 28 15:49:38 crc kubenswrapper[4805]: E1128 15:49:38.893775 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5\": container with ID starting with c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5 not found: ID does not exist" containerID="c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.893804 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5"} err="failed to get container status \"c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5\": rpc error: code = NotFound desc = could not find container \"c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5\": container with ID starting with c5000ba15bc5948ce287a2c3117a958c10da41ea01568ca2be7152b4397f8fc5 not found: ID does not exist" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.900612 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1db4967f-554e-4336-913a-fcebe420cf0e-kube-api-access-qmfd7" (OuterVolumeSpecName: "kube-api-access-qmfd7") pod "1db4967f-554e-4336-913a-fcebe420cf0e" (UID: "1db4967f-554e-4336-913a-fcebe420cf0e"). InnerVolumeSpecName "kube-api-access-qmfd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.909667 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ef9cc1-28ab-4014-897c-9679f3d36443-kube-api-access-vx5gn" (OuterVolumeSpecName: "kube-api-access-vx5gn") pod "19ef9cc1-28ab-4014-897c-9679f3d36443" (UID: "19ef9cc1-28ab-4014-897c-9679f3d36443"). InnerVolumeSpecName "kube-api-access-vx5gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.942293 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.972673 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-config-data" (OuterVolumeSpecName: "config-data") pod "d7348256-32de-4696-9637-96d7323831ed" (UID: "d7348256-32de-4696-9637-96d7323831ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.973210 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "19ef9cc1-28ab-4014-897c-9679f3d36443" (UID: "19ef9cc1-28ab-4014-897c-9679f3d36443"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.978852 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx5gn\" (UniqueName: \"kubernetes.io/projected/19ef9cc1-28ab-4014-897c-9679f3d36443-kube-api-access-vx5gn\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.978896 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.978911 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9mb4\" (UniqueName: \"kubernetes.io/projected/cf2e5369-a11b-4150-a291-4aeab9724d82-kube-api-access-r9mb4\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.978925 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db4967f-554e-4336-913a-fcebe420cf0e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.978936 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd6fbaf-5702-415e-8a1d-85d2c4792031-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.978961 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.978972 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x45zf\" (UniqueName: \"kubernetes.io/projected/d7348256-32de-4696-9637-96d7323831ed-kube-api-access-x45zf\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.978984 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf2e5369-a11b-4150-a291-4aeab9724d82-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.978995 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.979006 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2s4b\" (UniqueName: \"kubernetes.io/projected/ecd6fbaf-5702-415e-8a1d-85d2c4792031-kube-api-access-w2s4b\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.979017 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7348256-32de-4696-9637-96d7323831ed-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.979029 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmfd7\" (UniqueName: \"kubernetes.io/projected/1db4967f-554e-4336-913a-fcebe420cf0e-kube-api-access-qmfd7\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:38 crc kubenswrapper[4805]: I1128 15:49:38.979040 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ef9cc1-28ab-4014-897c-9679f3d36443-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.003524 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19ef9cc1-28ab-4014-897c-9679f3d36443" (UID: "19ef9cc1-28ab-4014-897c-9679f3d36443"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.017370 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7348256-32de-4696-9637-96d7323831ed" (UID: "d7348256-32de-4696-9637-96d7323831ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.028881 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.035719 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "19ef9cc1-28ab-4014-897c-9679f3d36443" (UID: "19ef9cc1-28ab-4014-897c-9679f3d36443"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.042012 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-config-data" (OuterVolumeSpecName: "config-data") pod "19ef9cc1-28ab-4014-897c-9679f3d36443" (UID: "19ef9cc1-28ab-4014-897c-9679f3d36443"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.043917 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementbd4b-account-delete-7bztm" event={"ID":"1db4967f-554e-4336-913a-fcebe420cf0e","Type":"ContainerDied","Data":"570c6cd900b4027822aaea393078090b0aa91ff67f91667fd3dafe88467fcc98"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.043951 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="570c6cd900b4027822aaea393078090b0aa91ff67f91667fd3dafe88467fcc98" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.044000 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementbd4b-account-delete-7bztm" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.053710 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.056801 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8232-account-delete-52zhn" event={"ID":"2849f80f-ed31-4c7f-8f65-1132aa35b6e3","Type":"ContainerStarted","Data":"ac4df6496bd374f34589a6d93fe9c479a53303f520c345b7a742bf81cedf5118"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.057479 4805 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican8232-account-delete-52zhn" secret="" err="secret \"galera-openstack-dockercfg-8kzms\" not found" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.058982 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d7348256-32de-4696-9637-96d7323831ed" (UID: "d7348256-32de-4696-9637-96d7323831ed"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.077296 4805 generic.go:334] "Generic (PLEG): container finished" podID="bd26b960-0994-4bb3-aef3-035519e32420" containerID="6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8" exitCode=0 Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.077455 4805 generic.go:334] "Generic (PLEG): container finished" podID="bd26b960-0994-4bb3-aef3-035519e32420" containerID="0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c" exitCode=0 Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.077538 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerDied","Data":"6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.077582 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerDied","Data":"0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.077737 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.082156 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg7wp\" (UniqueName: \"kubernetes.io/projected/1f9f5677-6d82-4ee0-bc2e-74c1671be521-kube-api-access-qg7wp\") pod \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\" (UID: \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\") " Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.082257 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9f5677-6d82-4ee0-bc2e-74c1671be521-operator-scripts\") pod \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\" (UID: \"1f9f5677-6d82-4ee0-bc2e-74c1671be521\") " Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.082951 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.082970 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.082982 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.082994 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.083005 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ef9cc1-28ab-4014-897c-9679f3d36443-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.083016 4805 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7348256-32de-4696-9637-96d7323831ed-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.084839 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f9f5677-6d82-4ee0-bc2e-74c1671be521-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1f9f5677-6d82-4ee0-bc2e-74c1671be521" (UID: "1f9f5677-6d82-4ee0-bc2e-74c1671be521"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.084992 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.085022 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-55d54b965d-bfqn6"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.085050 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts podName:2849f80f-ed31-4c7f-8f65-1132aa35b6e3 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:39.585030535 +0000 UTC m=+1406.634821846 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts") pod "barbican8232-account-delete-52zhn" (UID: "2849f80f-ed31-4c7f-8f65-1132aa35b6e3") : configmap "openstack-scripts" not found Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.086840 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56786c7f4d-vg252" event={"ID":"659e741d-dc3f-4463-b368-075d37a3ef0f","Type":"ContainerStarted","Data":"272a9c8516e1d98cd20d7f93294bdc7b48981088c857280f1f6c8fb621fcca83"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.086897 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.086900 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-56786c7f4d-vg252" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerName="barbican-api-log" containerID="cri-o://cbaaf427e82b90388021256c22b24bac4a5e83fbe17c6b89d768e25df50d48c0" gracePeriod=30 Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.086923 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-56786c7f4d-vg252" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerName="barbican-api" containerID="cri-o://272a9c8516e1d98cd20d7f93294bdc7b48981088c857280f1f6c8fb621fcca83" gracePeriod=30 Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.086916 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.087221 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f9f5677-6d82-4ee0-bc2e-74c1671be521-kube-api-access-qg7wp" (OuterVolumeSpecName: "kube-api-access-qg7wp") pod "1f9f5677-6d82-4ee0-bc2e-74c1671be521" (UID: "1f9f5677-6d82-4ee0-bc2e-74c1671be521"). InnerVolumeSpecName "kube-api-access-qg7wp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.094629 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.095990 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-55d54b965d-bfqn6"] Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.106480 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-97ccb4d45-ddn2r"] Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.121301 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7849b77bf-z2qp2" event={"ID":"95167e93-42b8-4f5f-b5a9-587b9b854f12","Type":"ContainerStarted","Data":"0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.121453 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7849b77bf-z2qp2" podUID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerName="barbican-worker-log" containerID="cri-o://3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf" gracePeriod=30 Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.121514 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7849b77bf-z2qp2" podUID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerName="barbican-worker" containerID="cri-o://0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c" gracePeriod=30 Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.133212 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.133477 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.135001 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.137657 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.137709 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.138639 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.140421 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-97ccb4d45-ddn2r"] Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.149579 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell08782-account-delete-phg59" event={"ID":"ecd6fbaf-5702-415e-8a1d-85d2c4792031","Type":"ContainerDied","Data":"54d1373449de9e63a701b72d93d3f049df833c6829ad0743cd08f91194871936"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.149618 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54d1373449de9e63a701b72d93d3f049df833c6829ad0743cd08f91194871936" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.149688 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell08782-account-delete-phg59" Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.158819 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.158882 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.164023 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican8232-account-delete-52zhn" podStartSLOduration=8.164006585 podStartE2EDuration="8.164006585s" podCreationTimestamp="2025-11-28 15:49:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:49:39.089664382 +0000 UTC m=+1406.139455703" watchObservedRunningTime="2025-11-28 15:49:39.164006585 +0000 UTC m=+1406.213797896" Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.170390 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.173187 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.174735 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.174774 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" containerName="nova-cell0-conductor-conductor" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.181264 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-56786c7f4d-vg252" podStartSLOduration=9.181247729 podStartE2EDuration="9.181247729s" podCreationTimestamp="2025-11-28 15:49:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:49:39.120903231 +0000 UTC m=+1406.170694552" watchObservedRunningTime="2025-11-28 15:49:39.181247729 +0000 UTC m=+1406.231039040" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.181484 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.181533 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"19ef9cc1-28ab-4014-897c-9679f3d36443","Type":"ContainerDied","Data":"93d4d2656c06ae1edfa165df51ed76d714acbc02dde56fdfceb090c51aeebe72"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.181576 4805 scope.go:117] "RemoveContainer" containerID="868c845c22044f953c99bf70b59a10a7cf406a71b3f9ce6c6262250153d75b30" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.183779 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd9hs\" (UniqueName: \"kubernetes.io/projected/f912081e-175b-410c-bbfa-daa3a71e2179-kube-api-access-xd9hs\") pod \"f912081e-175b-410c-bbfa-daa3a71e2179\" (UID: \"f912081e-175b-410c-bbfa-daa3a71e2179\") " Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.183938 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f912081e-175b-410c-bbfa-daa3a71e2179-operator-scripts\") pod \"f912081e-175b-410c-bbfa-daa3a71e2179\" (UID: \"f912081e-175b-410c-bbfa-daa3a71e2179\") " Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.183968 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svmdh\" (UniqueName: \"kubernetes.io/projected/9d7a9256-533a-4985-9a62-9bbe76af9e54-kube-api-access-svmdh\") pod \"9d7a9256-533a-4985-9a62-9bbe76af9e54\" (UID: \"9d7a9256-533a-4985-9a62-9bbe76af9e54\") " Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.183993 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d7a9256-533a-4985-9a62-9bbe76af9e54-operator-scripts\") pod \"9d7a9256-533a-4985-9a62-9bbe76af9e54\" (UID: \"9d7a9256-533a-4985-9a62-9bbe76af9e54\") " Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.184682 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg7wp\" (UniqueName: \"kubernetes.io/projected/1f9f5677-6d82-4ee0-bc2e-74c1671be521-kube-api-access-qg7wp\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.184709 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9f5677-6d82-4ee0-bc2e-74c1671be521-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.185294 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f912081e-175b-410c-bbfa-daa3a71e2179-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f912081e-175b-410c-bbfa-daa3a71e2179" (UID: "f912081e-175b-410c-bbfa-daa3a71e2179"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.185539 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d7a9256-533a-4985-9a62-9bbe76af9e54-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d7a9256-533a-4985-9a62-9bbe76af9e54" (UID: "9d7a9256-533a-4985-9a62-9bbe76af9e54"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.193024 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d7a9256-533a-4985-9a62-9bbe76af9e54-kube-api-access-svmdh" (OuterVolumeSpecName: "kube-api-access-svmdh") pod "9d7a9256-533a-4985-9a62-9bbe76af9e54" (UID: "9d7a9256-533a-4985-9a62-9bbe76af9e54"). InnerVolumeSpecName "kube-api-access-svmdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.195113 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance73a6-account-delete-qjlt7" event={"ID":"cf2e5369-a11b-4150-a291-4aeab9724d82","Type":"ContainerDied","Data":"7801d4719043577d7f80f1ec7425157934dda881fde1891589a1b9990a282b68"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.195154 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7801d4719043577d7f80f1ec7425157934dda881fde1891589a1b9990a282b68" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.195222 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance73a6-account-delete-qjlt7" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.204033 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f912081e-175b-410c-bbfa-daa3a71e2179-kube-api-access-xd9hs" (OuterVolumeSpecName: "kube-api-access-xd9hs") pod "f912081e-175b-410c-bbfa-daa3a71e2179" (UID: "f912081e-175b-410c-bbfa-daa3a71e2179"). InnerVolumeSpecName "kube-api-access-xd9hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.217100 4805 generic.go:334] "Generic (PLEG): container finished" podID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerID="42eba8d003a7a3309a588e3a7f4e6e3ccf9656400a9603ef4fb2b81bd3c138d4" exitCode=0 Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.239460 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" path="/var/lib/kubelet/pods/0e72cc62-edd5-4d0c-89a8-099d6b07e1e6/volumes" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.240402 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33489fb0-b55a-4544-8a36-c4e0c9ca10f0" path="/var/lib/kubelet/pods/33489fb0-b55a-4544-8a36-c4e0c9ca10f0/volumes" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.241056 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="452c347f-4ee3-46de-ba8f-c83300966f5d" path="/var/lib/kubelet/pods/452c347f-4ee3-46de-ba8f-c83300966f5d/volumes" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.243073 4805 generic.go:334] "Generic (PLEG): container finished" podID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerID="d1b2e011187c0efc47f70c8cd2cec54317c7da156f1da659b8207f714d6684f5" exitCode=0 Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.245404 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51493fc6-cc57-4bbf-a6a8-0864160ade3f" path="/var/lib/kubelet/pods/51493fc6-cc57-4bbf-a6a8-0864160ade3f/volumes" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.246295 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" path="/var/lib/kubelet/pods/6ef711de-53e6-4705-8d2f-469b6dc2d4de/volumes" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.248403 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abfd499c-e83b-4616-a80c-29a7e8e750e2" path="/var/lib/kubelet/pods/abfd499c-e83b-4616-a80c-29a7e8e750e2/volumes" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.249065 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b36ae016-8ef7-4c1f-a902-9e6f33d9aaad" path="/var/lib/kubelet/pods/b36ae016-8ef7-4c1f-a902-9e6f33d9aaad/volumes" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.251404 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" path="/var/lib/kubelet/pods/be5d0310-fb90-4d73-94f3-57b105a2a408/volumes" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.252410 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7849b77bf-z2qp2" podStartSLOduration=10.252393423000001 podStartE2EDuration="10.252393423s" podCreationTimestamp="2025-11-28 15:49:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:49:39.18529695 +0000 UTC m=+1406.235088281" watchObservedRunningTime="2025-11-28 15:49:39.252393423 +0000 UTC m=+1406.302184734" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.269615 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" event={"ID":"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6","Type":"ContainerDied","Data":"42eba8d003a7a3309a588e3a7f4e6e3ccf9656400a9603ef4fb2b81bd3c138d4"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.269964 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv5zc" event={"ID":"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5","Type":"ContainerDied","Data":"d1b2e011187c0efc47f70c8cd2cec54317c7da156f1da659b8207f714d6684f5"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.288622 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f912081e-175b-410c-bbfa-daa3a71e2179-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.288865 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svmdh\" (UniqueName: \"kubernetes.io/projected/9d7a9256-533a-4985-9a62-9bbe76af9e54-kube-api-access-svmdh\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.288931 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d7a9256-533a-4985-9a62-9bbe76af9e54-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.288990 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd9hs\" (UniqueName: \"kubernetes.io/projected/f912081e-175b-410c-bbfa-daa3a71e2179-kube-api-access-xd9hs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.302772 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7348256-32de-4696-9637-96d7323831ed","Type":"ContainerDied","Data":"75571db30208cbfdded485fc7f0b84911c3aaf764990bcced6068453ceb1e3da"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.302916 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.305203 4805 generic.go:334] "Generic (PLEG): container finished" podID="40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" containerID="01f4eeccdd9eae514a9fe127c422ee73a939f8ea3e4c0008f73fe1ea10f2ee26" exitCode=0 Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.305325 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.305465 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29","Type":"ContainerDied","Data":"01f4eeccdd9eae514a9fe127c422ee73a939f8ea3e4c0008f73fe1ea10f2ee26"} Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.305587 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.305643 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.306322 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.312456 4805 scope.go:117] "RemoveContainer" containerID="4d900ec2e1183b4ce27e8479bb7bc8fa6d050fc7702be965871220be9e57e825" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.379189 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell08782-account-delete-phg59"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.383130 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.384644 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.392291 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 28 15:49:39 crc kubenswrapper[4805]: E1128 15:49:39.392390 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="624edf72-de15-4026-812b-36d993917176" containerName="ovn-northd" Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.403441 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell08782-account-delete-phg59"] Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.412565 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:49:39 crc kubenswrapper[4805]: I1128 15:49:39.421828 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:39.602610 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:39.602672 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts podName:2849f80f-ed31-4c7f-8f65-1132aa35b6e3 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:40.602657297 +0000 UTC m=+1407.652448608 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts") pod "barbican8232-account-delete-52zhn" (UID: "2849f80f-ed31-4c7f-8f65-1132aa35b6e3") : configmap "openstack-scripts" not found Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.807297 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.812230 4805 scope.go:117] "RemoveContainer" containerID="5b7e15d112969a7b953c22fe09274e56dd365f74ce49e846d9443a14feb3ce7a" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.814736 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4q9c2" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" probeResult="failure" output="command timed out" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.822709 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.855426 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.856009 4805 scope.go:117] "RemoveContainer" containerID="099d02bc7f77ee26f120af8f3d1407c72c14067b902a1e28ada5fa2095d8e0a1" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.863901 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4q9c2" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" probeResult="failure" output=< Nov 28 15:49:40 crc kubenswrapper[4805]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Nov 28 15:49:40 crc kubenswrapper[4805]: > Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.885159 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.905559 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917197 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-erlang-cookie\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917256 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-scripts\") pod \"5c021413-239d-4ba0-979d-1c4b25d5093c\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917283 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-logs\") pod \"5c021413-239d-4ba0-979d-1c4b25d5093c\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917303 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmdz7\" (UniqueName: \"kubernetes.io/projected/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kube-api-access-nmdz7\") pod \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917333 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-plugins\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917378 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqvzt\" (UniqueName: \"kubernetes.io/projected/5c021413-239d-4ba0-979d-1c4b25d5093c-kube-api-access-dqvzt\") pod \"5c021413-239d-4ba0-979d-1c4b25d5093c\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917420 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-public-tls-certs\") pod \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917440 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-confd\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917470 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-public-tls-certs\") pod \"5c021413-239d-4ba0-979d-1c4b25d5093c\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917492 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-internal-tls-certs\") pod \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917538 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data\") pod \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917570 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-config-data\") pod \"5c021413-239d-4ba0-979d-1c4b25d5093c\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917600 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-combined-ca-bundle\") pod \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917620 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-memcached-tls-certs\") pod \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917641 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-server-conf\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917663 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-combined-ca-bundle\") pod \"5c021413-239d-4ba0-979d-1c4b25d5093c\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917692 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom\") pod \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917731 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjslc\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-kube-api-access-wjslc\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917754 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-erlang-cookie-secret\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917774 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-config-data\") pod \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917796 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917830 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-logs\") pod \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917871 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"5c021413-239d-4ba0-979d-1c4b25d5093c\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917912 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-pod-info\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917939 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-tls\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917980 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-httpd-run\") pod \"5c021413-239d-4ba0-979d-1c4b25d5093c\" (UID: \"5c021413-239d-4ba0-979d-1c4b25d5093c\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.917995 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.918027 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-combined-ca-bundle\") pod \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.918056 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kolla-config\") pod \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\" (UID: \"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.918076 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.918099 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-plugins-conf\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.918127 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksfft\" (UniqueName: \"kubernetes.io/projected/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-kube-api-access-ksfft\") pod \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\" (UID: \"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6\") " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.918584 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.921017 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.921022 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-logs" (OuterVolumeSpecName: "logs") pod "5c021413-239d-4ba0-979d-1c4b25d5093c" (UID: "5c021413-239d-4ba0-979d-1c4b25d5093c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.922960 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-kube-api-access-ksfft" (OuterVolumeSpecName: "kube-api-access-ksfft") pod "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6"). InnerVolumeSpecName "kube-api-access-ksfft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.923635 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" (UID: "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.924965 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.928723 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.931176 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-pod-info" (OuterVolumeSpecName: "pod-info") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.932153 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-scripts" (OuterVolumeSpecName: "scripts") pod "5c021413-239d-4ba0-979d-1c4b25d5093c" (UID: "5c021413-239d-4ba0-979d-1c4b25d5093c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.932238 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kube-api-access-nmdz7" (OuterVolumeSpecName: "kube-api-access-nmdz7") pod "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" (UID: "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29"). InnerVolumeSpecName "kube-api-access-nmdz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.937252 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-config-data" (OuterVolumeSpecName: "config-data") pod "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" (UID: "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.937565 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5c021413-239d-4ba0-979d-1c4b25d5093c" (UID: "5c021413-239d-4ba0-979d-1c4b25d5093c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.937575 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-logs" (OuterVolumeSpecName: "logs") pod "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.938131 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.941338 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.945062 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "5c021413-239d-4ba0-979d-1c4b25d5093c" (UID: "5c021413-239d-4ba0-979d-1c4b25d5093c"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.954633 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.958633 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.958755 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.958908 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-kube-api-access-wjslc" (OuterVolumeSpecName: "kube-api-access-wjslc") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "kube-api-access-wjslc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.965649 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c021413-239d-4ba0-979d-1c4b25d5093c-kube-api-access-dqvzt" (OuterVolumeSpecName: "kube-api-access-dqvzt") pod "5c021413-239d-4ba0-979d-1c4b25d5093c" (UID: "5c021413-239d-4ba0-979d-1c4b25d5093c"). InnerVolumeSpecName "kube-api-access-dqvzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.988670 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.995255 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-5c6d5f89c4-q6gr4"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:39.996639 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data" (OuterVolumeSpecName: "config-data") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.002333 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.012854 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.017655 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" (UID: "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019866 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjslc\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-kube-api-access-wjslc\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019888 4805 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019897 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019924 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019933 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019947 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019957 4805 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-pod-info\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019965 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019974 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019982 4805 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019989 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.019997 4805 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.020006 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksfft\" (UniqueName: \"kubernetes.io/projected/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-kube-api-access-ksfft\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.020014 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.020022 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.020029 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c021413-239d-4ba0-979d-1c4b25d5093c-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.020038 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmdz7\" (UniqueName: \"kubernetes.io/projected/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-kube-api-access-nmdz7\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.020046 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqvzt\" (UniqueName: \"kubernetes.io/projected/5c021413-239d-4ba0-979d-1c4b25d5093c-kube-api-access-dqvzt\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.020054 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.020062 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.021820 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.034638 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data" (OuterVolumeSpecName: "config-data") pod "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.040347 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.042256 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c021413-239d-4ba0-979d-1c4b25d5093c" (UID: "5c021413-239d-4ba0-979d-1c4b25d5093c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.049040 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.053699 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" (UID: "40e151fc-3a1e-4b10-8a6a-59bbb17b8d29"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.061562 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-config-data" (OuterVolumeSpecName: "config-data") pod "5c021413-239d-4ba0-979d-1c4b25d5093c" (UID: "5c021413-239d-4ba0-979d-1c4b25d5093c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.078307 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5c021413-239d-4ba0-979d-1c4b25d5093c" (UID: "5c021413-239d-4ba0-979d-1c4b25d5093c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.091056 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-server-conf" (OuterVolumeSpecName: "server-conf") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.093823 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.095488 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" (UID: "5c4171ee-19b0-44c4-8e98-d14e11b9e1a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.120803 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.120945 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-confd\") pod \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\" (UID: \"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a\") " Nov 28 15:49:40 crc kubenswrapper[4805]: W1128 15:49:40.121042 4805 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a/volumes/kubernetes.io~projected/rabbitmq-confd Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121067 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" (UID: "4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121383 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121399 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121410 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121422 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121431 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121441 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121450 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121460 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121469 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121479 4805 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a-server-conf\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121493 4805 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.121502 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c021413-239d-4ba0-979d-1c4b25d5093c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:40.121842 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:40.121958 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data podName:fb122aae-0e09-46b2-926c-037d25e79477 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:48.121930466 +0000 UTC m=+1415.171721777 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb122aae-0e09-46b2-926c-037d25e79477") : configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.324965 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic2b1-account-delete-wfngd" event={"ID":"9d7a9256-533a-4985-9a62-9bbe76af9e54","Type":"ContainerDied","Data":"47a72f32c1d827f7d0013132d7511e7b4f34f9dadbafbc8bbc216fa1f60da2a1"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.325408 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47a72f32c1d827f7d0013132d7511e7b4f34f9dadbafbc8bbc216fa1f60da2a1" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.325002 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic2b1-account-delete-wfngd" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.328490 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron1f9e-account-delete-vklns" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.328731 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron1f9e-account-delete-vklns" event={"ID":"1f9f5677-6d82-4ee0-bc2e-74c1671be521","Type":"ContainerDied","Data":"d81354992a8c0a17cb7d14a2b64856356fd98e98df5111476d121afdd30da0a4"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.328865 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d81354992a8c0a17cb7d14a2b64856356fd98e98df5111476d121afdd30da0a4" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.332257 4805 generic.go:334] "Generic (PLEG): container finished" podID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerID="cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9" exitCode=0 Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.332302 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a","Type":"ContainerDied","Data":"cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.332320 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a","Type":"ContainerDied","Data":"d553f54295d1ab3f97b6a308ebac8e0039b9aaecbf4c5e8b9ced6cffda3fc156"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.332337 4805 scope.go:117] "RemoveContainer" containerID="cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.332462 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.338727 4805 generic.go:334] "Generic (PLEG): container finished" podID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerID="272a9c8516e1d98cd20d7f93294bdc7b48981088c857280f1f6c8fb621fcca83" exitCode=0 Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.338815 4805 generic.go:334] "Generic (PLEG): container finished" podID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerID="cbaaf427e82b90388021256c22b24bac4a5e83fbe17c6b89d768e25df50d48c0" exitCode=143 Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.338783 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56786c7f4d-vg252" event={"ID":"659e741d-dc3f-4463-b368-075d37a3ef0f","Type":"ContainerDied","Data":"272a9c8516e1d98cd20d7f93294bdc7b48981088c857280f1f6c8fb621fcca83"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.338895 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56786c7f4d-vg252" event={"ID":"659e741d-dc3f-4463-b368-075d37a3ef0f","Type":"ContainerDied","Data":"cbaaf427e82b90388021256c22b24bac4a5e83fbe17c6b89d768e25df50d48c0"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.353297 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c021413-239d-4ba0-979d-1c4b25d5093c","Type":"ContainerDied","Data":"8492df5731462df2cd978bdbc1c0b2d7c30e2f8a9a1fa8380cd034b50491f736"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.353323 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.358429 4805 generic.go:334] "Generic (PLEG): container finished" podID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerID="3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf" exitCode=143 Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.358514 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7849b77bf-z2qp2" event={"ID":"95167e93-42b8-4f5f-b5a9-587b9b854f12","Type":"ContainerDied","Data":"3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.368627 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.369604 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fc4b9b6bd-9nsgk" event={"ID":"5c4171ee-19b0-44c4-8e98-d14e11b9e1a6","Type":"ContainerDied","Data":"bde01a0dd31be7ceea17360d55a6773739af8157b4ce25586760b7f62af80a26"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.376031 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.376174 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"40e151fc-3a1e-4b10-8a6a-59bbb17b8d29","Type":"ContainerDied","Data":"925b22ec00321cc888cc9fbf692470a9a21f58a9e31bdea13424e9a4e8bc0f0d"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.384970 4805 generic.go:334] "Generic (PLEG): container finished" podID="5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" containerID="d649033916f10c5d8b5e440164330092d201d1409da92cf8c71f690663343e7c" exitCode=0 Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.385041 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4","Type":"ContainerDied","Data":"d649033916f10c5d8b5e440164330092d201d1409da92cf8c71f690663343e7c"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.389253 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder72ad-account-delete-wnr85" event={"ID":"f912081e-175b-410c-bbfa-daa3a71e2179","Type":"ContainerDied","Data":"43ef2e24a7fed0cbd317e912ee73a5e94f395a6852f2a523cc0b0005666c7849"} Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.389293 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43ef2e24a7fed0cbd317e912ee73a5e94f395a6852f2a523cc0b0005666c7849" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.389306 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonefa91-account-delete-rcm8w" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.389346 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder72ad-account-delete-wnr85" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.397257 4805 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican8232-account-delete-52zhn" secret="" err="secret \"galera-openstack-dockercfg-8kzms\" not found" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.434848 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.441551 4805 scope.go:117] "RemoveContainer" containerID="bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.476998 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.500590 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/keystone-5f69cf9fcb-rvfkd" podUID="614a68ec-3129-413f-abb1-40a73ad9137e" containerName="keystone-api" probeResult="failure" output="Get \"https://10.217.0.148:5000/v3\": read tcp 10.217.0.2:49404->10.217.0.148:5000: read: connection reset by peer" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.502511 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.513024 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:40.523352 4805 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 28 15:49:40 crc kubenswrapper[4805]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-11-28T15:49:33Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 28 15:49:40 crc kubenswrapper[4805]: /etc/init.d/functions: line 589: 414 Alarm clock "$@" Nov 28 15:49:40 crc kubenswrapper[4805]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-4q9c2" message=< Nov 28 15:49:40 crc kubenswrapper[4805]: Exiting ovn-controller (1) [FAILED] Nov 28 15:49:40 crc kubenswrapper[4805]: Killing ovn-controller (1) [ OK ] Nov 28 15:49:40 crc kubenswrapper[4805]: Killing ovn-controller (1) with SIGKILL [ OK ] Nov 28 15:49:40 crc kubenswrapper[4805]: 2025-11-28T15:49:33Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 28 15:49:40 crc kubenswrapper[4805]: /etc/init.d/functions: line 589: 414 Alarm clock "$@" Nov 28 15:49:40 crc kubenswrapper[4805]: > Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:40.523433 4805 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 28 15:49:40 crc kubenswrapper[4805]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-11-28T15:49:33Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 28 15:49:40 crc kubenswrapper[4805]: /etc/init.d/functions: line 589: 414 Alarm clock "$@" Nov 28 15:49:40 crc kubenswrapper[4805]: > pod="openstack/ovn-controller-4q9c2" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" containerID="cri-o://fff36b28f1a692a162fd59eb5953068bcc570d89664204a8c1cd194a602cd634" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.523465 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-4q9c2" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" containerID="cri-o://fff36b28f1a692a162fd59eb5953068bcc570d89664204a8c1cd194a602cd634" gracePeriod=22 Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.530899 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5fc4b9b6bd-9nsgk"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.541348 4805 scope.go:117] "RemoveContainer" containerID="cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9" Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:40.542742 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9\": container with ID starting with cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9 not found: ID does not exist" containerID="cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.542772 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9"} err="failed to get container status \"cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9\": rpc error: code = NotFound desc = could not find container \"cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9\": container with ID starting with cf36e60268aadda65af38beaa6932c386b6962ec43e90d40359c952acc28d8b9 not found: ID does not exist" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.542790 4805 scope.go:117] "RemoveContainer" containerID="bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.543766 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5fc4b9b6bd-9nsgk"] Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:40.546075 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52\": container with ID starting with bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52 not found: ID does not exist" containerID="bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.546124 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52"} err="failed to get container status \"bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52\": rpc error: code = NotFound desc = could not find container \"bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52\": container with ID starting with bb160967be02e9fcad7a1724158891a58220689d68dd4e513dd1deb3d4cb9e52 not found: ID does not exist" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.546147 4805 scope.go:117] "RemoveContainer" containerID="31105a43fdab9daa89f1ca09a13406cb4b8904817a89036a2573227173287ae5" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.554544 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.560985 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.573431 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystonefa91-account-delete-rcm8w"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.579785 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystonefa91-account-delete-rcm8w"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.645808 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ntr2\" (UniqueName: \"kubernetes.io/projected/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-kube-api-access-5ntr2\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.645844 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:40.645918 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:49:40 crc kubenswrapper[4805]: E1128 15:49:40.646001 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts podName:2849f80f-ed31-4c7f-8f65-1132aa35b6e3 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:42.645980345 +0000 UTC m=+1409.695771656 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts") pod "barbican8232-account-delete-52zhn" (UID: "2849f80f-ed31-4c7f-8f65-1132aa35b6e3") : configmap "openstack-scripts" not found Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.725200 4805 scope.go:117] "RemoveContainer" containerID="5290a7c9695a774d3cd36d80cc41c5b6f08e321f90789c87b7e2f50f6f5d2099" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.747121 4805 scope.go:117] "RemoveContainer" containerID="42eba8d003a7a3309a588e3a7f4e6e3ccf9656400a9603ef4fb2b81bd3c138d4" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.878659 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-f5hv2"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.911204 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-f5hv2"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.924028 4805 scope.go:117] "RemoveContainer" containerID="e4e089980980ac67dc67f214af92d43970eb148b6a828b8e14ba26fab9d1df86" Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.943228 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance73a6-account-delete-qjlt7"] Nov 28 15:49:40 crc kubenswrapper[4805]: I1128 15:49:40.977497 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-73a6-account-create-update-nfdpv"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.002427 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance73a6-account-delete-qjlt7"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.019625 4805 scope.go:117] "RemoveContainer" containerID="01f4eeccdd9eae514a9fe127c422ee73a939f8ea3e4c0008f73fe1ea10f2ee26" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.059921 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-73a6-account-create-update-nfdpv"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.119490 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-jfqsr"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.164531 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-jfqsr"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.184709 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder72ad-account-delete-wnr85"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.219811 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19093324-6f6b-47d3-a08d-051d59136892" path="/var/lib/kubelet/pods/19093324-6f6b-47d3-a08d-051d59136892/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.220608 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ef9cc1-28ab-4014-897c-9679f3d36443" path="/var/lib/kubelet/pods/19ef9cc1-28ab-4014-897c-9679f3d36443/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.221204 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" path="/var/lib/kubelet/pods/40e151fc-3a1e-4b10-8a6a-59bbb17b8d29/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.223119 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" path="/var/lib/kubelet/pods/4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.223829 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ff5237a-ad35-4875-9cd7-046cfbef7506" path="/var/lib/kubelet/pods/4ff5237a-ad35-4875-9cd7-046cfbef7506/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.224790 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c021413-239d-4ba0-979d-1c4b25d5093c" path="/var/lib/kubelet/pods/5c021413-239d-4ba0-979d-1c4b25d5093c/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.225857 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" path="/var/lib/kubelet/pods/5c4171ee-19b0-44c4-8e98-d14e11b9e1a6/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.226895 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89d423d2-d06f-4c64-9312-2258d0a746a6" path="/var/lib/kubelet/pods/89d423d2-d06f-4c64-9312-2258d0a746a6/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.228815 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a363e2d1-0a91-4578-b9e4-b736b7931b03" path="/var/lib/kubelet/pods/a363e2d1-0a91-4578-b9e4-b736b7931b03/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.239258 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7" path="/var/lib/kubelet/pods/aef1c4c9-c8d8-40f0-ae26-9709a0e9fdb7/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.239684 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aefb91a4-432a-4c5d-8d28-dee413c660c3" path="/var/lib/kubelet/pods/aefb91a4-432a-4c5d-8d28-dee413c660c3/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.240154 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf2e5369-a11b-4150-a291-4aeab9724d82" path="/var/lib/kubelet/pods/cf2e5369-a11b-4150-a291-4aeab9724d82/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.244872 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7348256-32de-4696-9637-96d7323831ed" path="/var/lib/kubelet/pods/d7348256-32de-4696-9637-96d7323831ed/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.245072 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.245634 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd6fbaf-5702-415e-8a1d-85d2c4792031" path="/var/lib/kubelet/pods/ecd6fbaf-5702-415e-8a1d-85d2c4792031/volumes" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.246150 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder72ad-account-delete-wnr85"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.246175 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-72ad-account-create-update-66ckj"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.260496 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-72ad-account-create-update-66ckj"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.277921 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.294265 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-ph6gw"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.319795 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-ph6gw"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.324537 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_624edf72-de15-4026-812b-36d993917176/ovn-northd/0.log" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.324610 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.358121 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-bd4b-account-create-update-zmxqb"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.367760 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659e741d-dc3f-4463-b368-075d37a3ef0f-logs\") pod \"659e741d-dc3f-4463-b368-075d37a3ef0f\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.367826 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-combined-ca-bundle\") pod \"659e741d-dc3f-4463-b368-075d37a3ef0f\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.367871 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data-custom\") pod \"659e741d-dc3f-4463-b368-075d37a3ef0f\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.367933 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-internal-tls-certs\") pod \"659e741d-dc3f-4463-b368-075d37a3ef0f\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.367972 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data\") pod \"659e741d-dc3f-4463-b368-075d37a3ef0f\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.368038 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk5tw\" (UniqueName: \"kubernetes.io/projected/659e741d-dc3f-4463-b368-075d37a3ef0f-kube-api-access-qk5tw\") pod \"659e741d-dc3f-4463-b368-075d37a3ef0f\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.368139 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-public-tls-certs\") pod \"659e741d-dc3f-4463-b368-075d37a3ef0f\" (UID: \"659e741d-dc3f-4463-b368-075d37a3ef0f\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.369858 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/659e741d-dc3f-4463-b368-075d37a3ef0f-logs" (OuterVolumeSpecName: "logs") pod "659e741d-dc3f-4463-b368-075d37a3ef0f" (UID: "659e741d-dc3f-4463-b368-075d37a3ef0f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.383568 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "659e741d-dc3f-4463-b368-075d37a3ef0f" (UID: "659e741d-dc3f-4463-b368-075d37a3ef0f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.383594 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/659e741d-dc3f-4463-b368-075d37a3ef0f-kube-api-access-qk5tw" (OuterVolumeSpecName: "kube-api-access-qk5tw") pod "659e741d-dc3f-4463-b368-075d37a3ef0f" (UID: "659e741d-dc3f-4463-b368-075d37a3ef0f"). InnerVolumeSpecName "kube-api-access-qk5tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.393176 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-bd4b-account-create-update-zmxqb"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.401680 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56786c7f4d-vg252" event={"ID":"659e741d-dc3f-4463-b368-075d37a3ef0f","Type":"ContainerDied","Data":"bba6a6039d9c7d4ae4bc9035b58edee1014582a984d98d9604b06f7f1ef23f2c"} Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.401847 4805 scope.go:117] "RemoveContainer" containerID="272a9c8516e1d98cd20d7f93294bdc7b48981088c857280f1f6c8fb621fcca83" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.401945 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56786c7f4d-vg252" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.401803 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "659e741d-dc3f-4463-b368-075d37a3ef0f" (UID: "659e741d-dc3f-4463-b368-075d37a3ef0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.408111 4805 generic.go:334] "Generic (PLEG): container finished" podID="614a68ec-3129-413f-abb1-40a73ad9137e" containerID="28e9827ec5a663c00f28c55d8c943235e37131e6dc45f5d050381a2a732dd392" exitCode=0 Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.408216 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f69cf9fcb-rvfkd" event={"ID":"614a68ec-3129-413f-abb1-40a73ad9137e","Type":"ContainerDied","Data":"28e9827ec5a663c00f28c55d8c943235e37131e6dc45f5d050381a2a732dd392"} Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.409594 4805 generic.go:334] "Generic (PLEG): container finished" podID="4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" containerID="3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339" exitCode=0 Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.409691 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1","Type":"ContainerDied","Data":"3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339"} Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.411799 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4","Type":"ContainerDied","Data":"ef34273fa04965fe17f2bbd3aa0bf4f1851554d8ea6815f2445e709987a916ff"} Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.411895 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.422904 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_624edf72-de15-4026-812b-36d993917176/ovn-northd/0.log" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.422950 4805 generic.go:334] "Generic (PLEG): container finished" podID="624edf72-de15-4026-812b-36d993917176" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" exitCode=139 Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.422998 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"624edf72-de15-4026-812b-36d993917176","Type":"ContainerDied","Data":"10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d"} Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.423023 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"624edf72-de15-4026-812b-36d993917176","Type":"ContainerDied","Data":"9f01bde7a205b705624751896b1fa4f2374f5206679f5c472a61b41559d0d9b2"} Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.423082 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.423299 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "659e741d-dc3f-4463-b368-075d37a3ef0f" (UID: "659e741d-dc3f-4463-b368-075d37a3ef0f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.428207 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4q9c2_cffd57cb-2509-4a9e-8e5d-1750e4b0493e/ovn-controller/0.log" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.428254 4805 generic.go:334] "Generic (PLEG): container finished" podID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerID="fff36b28f1a692a162fd59eb5953068bcc570d89664204a8c1cd194a602cd634" exitCode=137 Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.428910 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2" event={"ID":"cffd57cb-2509-4a9e-8e5d-1750e4b0493e","Type":"ContainerDied","Data":"fff36b28f1a692a162fd59eb5953068bcc570d89664204a8c1cd194a602cd634"} Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.431901 4805 scope.go:117] "RemoveContainer" containerID="cbaaf427e82b90388021256c22b24bac4a5e83fbe17c6b89d768e25df50d48c0" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.440148 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv5zc" event={"ID":"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5","Type":"ContainerStarted","Data":"2769fb70dd932ad97b359e1ae451227e6a8973823113353f6d855fb56e16ed14"} Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.441432 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "659e741d-dc3f-4463-b368-075d37a3ef0f" (UID: "659e741d-dc3f-4463-b368-075d37a3ef0f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.461300 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementbd4b-account-delete-7bztm"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.463693 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data" (OuterVolumeSpecName: "config-data") pod "659e741d-dc3f-4463-b368-075d37a3ef0f" (UID: "659e741d-dc3f-4463-b368-075d37a3ef0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470020 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-galera-tls-certs\") pod \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470077 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470099 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/624edf72-de15-4026-812b-36d993917176-ovn-rundir\") pod \"624edf72-de15-4026-812b-36d993917176\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470123 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-scripts\") pod \"624edf72-de15-4026-812b-36d993917176\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470149 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-generated\") pod \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470175 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-default\") pod \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470203 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-config\") pod \"624edf72-de15-4026-812b-36d993917176\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470223 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kolla-config\") pod \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470243 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-metrics-certs-tls-certs\") pod \"624edf72-de15-4026-812b-36d993917176\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470266 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-ovn-northd-tls-certs\") pod \"624edf72-de15-4026-812b-36d993917176\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470295 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-combined-ca-bundle\") pod \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470312 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-operator-scripts\") pod \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470341 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-combined-ca-bundle\") pod \"624edf72-de15-4026-812b-36d993917176\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470385 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rwzl\" (UniqueName: \"kubernetes.io/projected/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kube-api-access-6rwzl\") pod \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\" (UID: \"5243b85f-7ec0-4cca-9f10-de8b40b0a0f4\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470477 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sfzs\" (UniqueName: \"kubernetes.io/projected/624edf72-de15-4026-812b-36d993917176-kube-api-access-2sfzs\") pod \"624edf72-de15-4026-812b-36d993917176\" (UID: \"624edf72-de15-4026-812b-36d993917176\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470847 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470860 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470870 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470879 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk5tw\" (UniqueName: \"kubernetes.io/projected/659e741d-dc3f-4463-b368-075d37a3ef0f-kube-api-access-qk5tw\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470889 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470897 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659e741d-dc3f-4463-b368-075d37a3ef0f-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.470907 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659e741d-dc3f-4463-b368-075d37a3ef0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.479413 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/624edf72-de15-4026-812b-36d993917176-kube-api-access-2sfzs" (OuterVolumeSpecName: "kube-api-access-2sfzs") pod "624edf72-de15-4026-812b-36d993917176" (UID: "624edf72-de15-4026-812b-36d993917176"). InnerVolumeSpecName "kube-api-access-2sfzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.483233 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" (UID: "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.485418 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" (UID: "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.487529 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/624edf72-de15-4026-812b-36d993917176-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "624edf72-de15-4026-812b-36d993917176" (UID: "624edf72-de15-4026-812b-36d993917176"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.487922 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-scripts" (OuterVolumeSpecName: "scripts") pod "624edf72-de15-4026-812b-36d993917176" (UID: "624edf72-de15-4026-812b-36d993917176"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.489661 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" (UID: "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.496313 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-config" (OuterVolumeSpecName: "config") pod "624edf72-de15-4026-812b-36d993917176" (UID: "624edf72-de15-4026-812b-36d993917176"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.497205 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" (UID: "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.508989 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kube-api-access-6rwzl" (OuterVolumeSpecName: "kube-api-access-6rwzl") pod "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" (UID: "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4"). InnerVolumeSpecName "kube-api-access-6rwzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.512273 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementbd4b-account-delete-7bztm"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.520022 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-bzx4d"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.522475 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "mysql-db") pod "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" (UID: "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.555449 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-bzx4d"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.565417 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron1f9e-account-delete-vklns"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.565778 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" (UID: "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.575928 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rwzl\" (UniqueName: \"kubernetes.io/projected/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kube-api-access-6rwzl\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.575951 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sfzs\" (UniqueName: \"kubernetes.io/projected/624edf72-de15-4026-812b-36d993917176-kube-api-access-2sfzs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.575974 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.575983 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/624edf72-de15-4026-812b-36d993917176-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.575992 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.576001 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.576011 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.576019 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624edf72-de15-4026-812b-36d993917176-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.576027 4805 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.576034 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.576042 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.584194 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "624edf72-de15-4026-812b-36d993917176" (UID: "624edf72-de15-4026-812b-36d993917176"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.584414 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-1f9e-account-create-update-c5tc4"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.604064 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.604491 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-1f9e-account-create-update-c5tc4"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.618669 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "624edf72-de15-4026-812b-36d993917176" (UID: "624edf72-de15-4026-812b-36d993917176"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.624548 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron1f9e-account-delete-vklns"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.650612 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "624edf72-de15-4026-812b-36d993917176" (UID: "624edf72-de15-4026-812b-36d993917176"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.670683 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" (UID: "5243b85f-7ec0-4cca-9f10-de8b40b0a0f4"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.673809 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-dnbhb"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.678111 4805 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.678145 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.678158 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.678172 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.678184 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624edf72-de15-4026-812b-36d993917176-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.734064 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-dnbhb"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.748158 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8232-account-delete-52zhn"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.754559 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican8232-account-delete-52zhn" podUID="2849f80f-ed31-4c7f-8f65-1132aa35b6e3" containerName="mariadb-account-delete" containerID="cri-o://ac4df6496bd374f34589a6d93fe9c479a53303f520c345b7a742bf81cedf5118" gracePeriod=30 Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.756790 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8232-account-create-update-tg9lf"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.764564 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8232-account-create-update-tg9lf"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.789970 4805 scope.go:117] "RemoveContainer" containerID="d649033916f10c5d8b5e440164330092d201d1409da92cf8c71f690663343e7c" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.791304 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-rr5bn"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.793259 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.818955 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4q9c2_cffd57cb-2509-4a9e-8e5d-1750e4b0493e/ovn-controller/0.log" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.819289 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.828495 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.829468 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-rr5bn"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.837088 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapic2b1-account-delete-wfngd"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.841602 4805 scope.go:117] "RemoveContainer" containerID="5e976116e68f4c0f2b31a953105e7f91891d03875fa4ded4f3ff8f472c7e6772" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.845395 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapic2b1-account-delete-wfngd"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.851478 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c2b1-account-create-update-qgsb8"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.857980 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c2b1-account-create-update-qgsb8"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.863964 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-56786c7f4d-vg252"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.870181 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-56786c7f4d-vg252"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.877416 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.885572 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7nvm\" (UniqueName: \"kubernetes.io/projected/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-kube-api-access-g7nvm\") pod \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.885607 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-combined-ca-bundle\") pod \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.885643 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-config-data\") pod \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\" (UID: \"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1\") " Nov 28 15:49:41 crc kubenswrapper[4805]: E1128 15:49:41.886138 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:41 crc kubenswrapper[4805]: E1128 15:49:41.886182 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:49.88616833 +0000 UTC m=+1416.935959641 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-config-data" not found Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.887166 4805 scope.go:117] "RemoveContainer" containerID="fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.887312 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.891923 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 15:49:41 crc kubenswrapper[4805]: E1128 15:49:41.896825 4805 secret.go:188] Couldn't get secret openstack/barbican-keystone-listener-config-data: secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:41 crc kubenswrapper[4805]: E1128 15:49:41.896907 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:49.896885474 +0000 UTC m=+1416.946676785 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.899662 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.900979 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-kube-api-access-g7nvm" (OuterVolumeSpecName: "kube-api-access-g7nvm") pod "4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" (UID: "4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1"). InnerVolumeSpecName "kube-api-access-g7nvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.930023 4805 scope.go:117] "RemoveContainer" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.931465 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" (UID: "4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.939191 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-config-data" (OuterVolumeSpecName: "config-data") pod "4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" (UID: "4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.948175 4805 scope.go:117] "RemoveContainer" containerID="fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574" Nov 28 15:49:41 crc kubenswrapper[4805]: E1128 15:49:41.950015 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574\": container with ID starting with fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574 not found: ID does not exist" containerID="fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.950079 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574"} err="failed to get container status \"fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574\": rpc error: code = NotFound desc = could not find container \"fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574\": container with ID starting with fc55319b61b431891847adb6f0e233232cf622a30d8d7c1b603fd252cd6ae574 not found: ID does not exist" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.950107 4805 scope.go:117] "RemoveContainer" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" Nov 28 15:49:41 crc kubenswrapper[4805]: E1128 15:49:41.950562 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d\": container with ID starting with 10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d not found: ID does not exist" containerID="10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.950593 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d"} err="failed to get container status \"10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d\": rpc error: code = NotFound desc = could not find container \"10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d\": container with ID starting with 10bbdaa35ce8e318d56dc77b88272908067e00d4182afc468d050c94feca566d not found: ID does not exist" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987348 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-log-ovn\") pod \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987437 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-public-tls-certs\") pod \"614a68ec-3129-413f-abb1-40a73ad9137e\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987492 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-scripts\") pod \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987520 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-combined-ca-bundle\") pod \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987541 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-ovn-controller-tls-certs\") pod \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987569 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-config-data\") pod \"614a68ec-3129-413f-abb1-40a73ad9137e\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987612 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xpv2\" (UniqueName: \"kubernetes.io/projected/614a68ec-3129-413f-abb1-40a73ad9137e-kube-api-access-2xpv2\") pod \"614a68ec-3129-413f-abb1-40a73ad9137e\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987648 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run\") pod \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987692 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run-ovn\") pod \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987726 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-combined-ca-bundle\") pod \"614a68ec-3129-413f-abb1-40a73ad9137e\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987759 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-fernet-keys\") pod \"614a68ec-3129-413f-abb1-40a73ad9137e\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987778 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-scripts\") pod \"614a68ec-3129-413f-abb1-40a73ad9137e\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987812 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjhkb\" (UniqueName: \"kubernetes.io/projected/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-kube-api-access-jjhkb\") pod \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\" (UID: \"cffd57cb-2509-4a9e-8e5d-1750e4b0493e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987827 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-internal-tls-certs\") pod \"614a68ec-3129-413f-abb1-40a73ad9137e\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.987848 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-credential-keys\") pod \"614a68ec-3129-413f-abb1-40a73ad9137e\" (UID: \"614a68ec-3129-413f-abb1-40a73ad9137e\") " Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.988217 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.988237 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7nvm\" (UniqueName: \"kubernetes.io/projected/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-kube-api-access-g7nvm\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.988247 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.989237 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "cffd57cb-2509-4a9e-8e5d-1750e4b0493e" (UID: "cffd57cb-2509-4a9e-8e5d-1750e4b0493e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.990431 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "cffd57cb-2509-4a9e-8e5d-1750e4b0493e" (UID: "cffd57cb-2509-4a9e-8e5d-1750e4b0493e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.990545 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run" (OuterVolumeSpecName: "var-run") pod "cffd57cb-2509-4a9e-8e5d-1750e4b0493e" (UID: "cffd57cb-2509-4a9e-8e5d-1750e4b0493e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.991600 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-scripts" (OuterVolumeSpecName: "scripts") pod "cffd57cb-2509-4a9e-8e5d-1750e4b0493e" (UID: "cffd57cb-2509-4a9e-8e5d-1750e4b0493e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.993261 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "614a68ec-3129-413f-abb1-40a73ad9137e" (UID: "614a68ec-3129-413f-abb1-40a73ad9137e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.996247 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "614a68ec-3129-413f-abb1-40a73ad9137e" (UID: "614a68ec-3129-413f-abb1-40a73ad9137e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:41 crc kubenswrapper[4805]: I1128 15:49:41.998015 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-scripts" (OuterVolumeSpecName: "scripts") pod "614a68ec-3129-413f-abb1-40a73ad9137e" (UID: "614a68ec-3129-413f-abb1-40a73ad9137e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.011887 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/614a68ec-3129-413f-abb1-40a73ad9137e-kube-api-access-2xpv2" (OuterVolumeSpecName: "kube-api-access-2xpv2") pod "614a68ec-3129-413f-abb1-40a73ad9137e" (UID: "614a68ec-3129-413f-abb1-40a73ad9137e"). InnerVolumeSpecName "kube-api-access-2xpv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.011948 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-kube-api-access-jjhkb" (OuterVolumeSpecName: "kube-api-access-jjhkb") pod "cffd57cb-2509-4a9e-8e5d-1750e4b0493e" (UID: "cffd57cb-2509-4a9e-8e5d-1750e4b0493e"). InnerVolumeSpecName "kube-api-access-jjhkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.015757 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-config-data" (OuterVolumeSpecName: "config-data") pod "614a68ec-3129-413f-abb1-40a73ad9137e" (UID: "614a68ec-3129-413f-abb1-40a73ad9137e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.025410 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cffd57cb-2509-4a9e-8e5d-1750e4b0493e" (UID: "cffd57cb-2509-4a9e-8e5d-1750e4b0493e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.035424 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "614a68ec-3129-413f-abb1-40a73ad9137e" (UID: "614a68ec-3129-413f-abb1-40a73ad9137e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.036397 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "614a68ec-3129-413f-abb1-40a73ad9137e" (UID: "614a68ec-3129-413f-abb1-40a73ad9137e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.057157 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "614a68ec-3129-413f-abb1-40a73ad9137e" (UID: "614a68ec-3129-413f-abb1-40a73ad9137e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.068291 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "cffd57cb-2509-4a9e-8e5d-1750e4b0493e" (UID: "cffd57cb-2509-4a9e-8e5d-1750e4b0493e"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089787 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xpv2\" (UniqueName: \"kubernetes.io/projected/614a68ec-3129-413f-abb1-40a73ad9137e-kube-api-access-2xpv2\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089832 4805 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089846 4805 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089885 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089901 4805 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089912 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089923 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjhkb\" (UniqueName: \"kubernetes.io/projected/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-kube-api-access-jjhkb\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089935 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089946 4805 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089957 4805 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089967 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089979 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.089990 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.090001 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cffd57cb-2509-4a9e-8e5d-1750e4b0493e-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.090011 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614a68ec-3129-413f-abb1-40a73ad9137e-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.482421 4805 generic.go:334] "Generic (PLEG): container finished" podID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerID="2769fb70dd932ad97b359e1ae451227e6a8973823113353f6d855fb56e16ed14" exitCode=0 Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.483740 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv5zc" event={"ID":"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5","Type":"ContainerDied","Data":"2769fb70dd932ad97b359e1ae451227e6a8973823113353f6d855fb56e16ed14"} Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.487683 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f69cf9fcb-rvfkd" event={"ID":"614a68ec-3129-413f-abb1-40a73ad9137e","Type":"ContainerDied","Data":"5e0db73f65439bf234434ee8cbf8f01e05605fc1ebe5e373a0f6994435fb3c43"} Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.487720 4805 scope.go:117] "RemoveContainer" containerID="28e9827ec5a663c00f28c55d8c943235e37131e6dc45f5d050381a2a732dd392" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.487929 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f69cf9fcb-rvfkd" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.493607 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.493632 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1","Type":"ContainerDied","Data":"82fb7f71741c2e2edd378bc77ad4fdff334b437d92ae7263f66e56dccca3ee9e"} Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.499236 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4q9c2_cffd57cb-2509-4a9e-8e5d-1750e4b0493e/ovn-controller/0.log" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.499296 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4q9c2" event={"ID":"cffd57cb-2509-4a9e-8e5d-1750e4b0493e","Type":"ContainerDied","Data":"a3b568960d9f9ff1ec39f0c731def0ddc8fa8bd09f13a349d65132b889f0a830"} Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.499374 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4q9c2" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.525735 4805 scope.go:117] "RemoveContainer" containerID="3633c2201e2b4a484a9a766779d6a1025519dc378ea68474da791a9f61ddd339" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.536087 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5f69cf9fcb-rvfkd"] Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.543729 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5f69cf9fcb-rvfkd"] Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.560419 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.566567 4805 scope.go:117] "RemoveContainer" containerID="fff36b28f1a692a162fd59eb5953068bcc570d89664204a8c1cd194a602cd634" Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.570382 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.576702 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4q9c2"] Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.582306 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4q9c2"] Nov 28 15:49:42 crc kubenswrapper[4805]: E1128 15:49:42.704178 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:49:42 crc kubenswrapper[4805]: E1128 15:49:42.704299 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts podName:2849f80f-ed31-4c7f-8f65-1132aa35b6e3 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:46.704276868 +0000 UTC m=+1413.754068179 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts") pod "barbican8232-account-delete-52zhn" (UID: "2849f80f-ed31-4c7f-8f65-1132aa35b6e3") : configmap "openstack-scripts" not found Nov 28 15:49:42 crc kubenswrapper[4805]: I1128 15:49:42.708171 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.163:8776/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.216128 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09c6b57a-16a1-465d-b883-aa6bea6756c5" path="/var/lib/kubelet/pods/09c6b57a-16a1-465d-b883-aa6bea6756c5/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.217037 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c22dd95-b8ab-44d5-8783-73ece53f5dc8" path="/var/lib/kubelet/pods/0c22dd95-b8ab-44d5-8783-73ece53f5dc8/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.217554 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1db4967f-554e-4336-913a-fcebe420cf0e" path="/var/lib/kubelet/pods/1db4967f-554e-4336-913a-fcebe420cf0e/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.218024 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f9f5677-6d82-4ee0-bc2e-74c1671be521" path="/var/lib/kubelet/pods/1f9f5677-6d82-4ee0-bc2e-74c1671be521/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.218929 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3672f82d-1221-47a2-8be9-cbc3394def33" path="/var/lib/kubelet/pods/3672f82d-1221-47a2-8be9-cbc3394def33/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.219656 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c642d47-f1ea-4254-a76e-5ebe959da352" path="/var/lib/kubelet/pods/3c642d47-f1ea-4254-a76e-5ebe959da352/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.220119 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" path="/var/lib/kubelet/pods/4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.220717 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" path="/var/lib/kubelet/pods/5243b85f-7ec0-4cca-9f10-de8b40b0a0f4/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.221740 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="614a68ec-3129-413f-abb1-40a73ad9137e" path="/var/lib/kubelet/pods/614a68ec-3129-413f-abb1-40a73ad9137e/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.222348 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="624edf72-de15-4026-812b-36d993917176" path="/var/lib/kubelet/pods/624edf72-de15-4026-812b-36d993917176/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.223000 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" path="/var/lib/kubelet/pods/659e741d-dc3f-4463-b368-075d37a3ef0f/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.224071 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a29bf56-ccc6-4b62-a4b7-da0093d80478" path="/var/lib/kubelet/pods/7a29bf56-ccc6-4b62-a4b7-da0093d80478/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.224554 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d7a9256-533a-4985-9a62-9bbe76af9e54" path="/var/lib/kubelet/pods/9d7a9256-533a-4985-9a62-9bbe76af9e54/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.225109 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c308b8e2-19af-4cfc-9a06-487497e95198" path="/var/lib/kubelet/pods/c308b8e2-19af-4cfc-9a06-487497e95198/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.226046 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" path="/var/lib/kubelet/pods/cffd57cb-2509-4a9e-8e5d-1750e4b0493e/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.226851 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6cb1991-b179-4169-be04-b0e707cc99a3" path="/var/lib/kubelet/pods/d6cb1991-b179-4169-be04-b0e707cc99a3/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.227338 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e937c330-5ffb-48ee-8607-fa0764ef7172" path="/var/lib/kubelet/pods/e937c330-5ffb-48ee-8607-fa0764ef7172/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.228540 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f912081e-175b-410c-bbfa-daa3a71e2179" path="/var/lib/kubelet/pods/f912081e-175b-410c-bbfa-daa3a71e2179/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.229040 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce" path="/var/lib/kubelet/pods/ffeedc8c-a5ef-4b3a-9094-9fdd1337e9ce/volumes" Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.512890 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv5zc" event={"ID":"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5","Type":"ContainerStarted","Data":"553af6e1e47c786d07950057537bac41ba4e9ef9d1b92cfd40080a325fe92c1e"} Nov 28 15:49:43 crc kubenswrapper[4805]: I1128 15:49:43.542996 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nv5zc" podStartSLOduration=7.707358554 podStartE2EDuration="11.542975623s" podCreationTimestamp="2025-11-28 15:49:32 +0000 UTC" firstStartedPulling="2025-11-28 15:49:39.257695739 +0000 UTC m=+1406.307487050" lastFinishedPulling="2025-11-28 15:49:43.093312808 +0000 UTC m=+1410.143104119" observedRunningTime="2025-11-28 15:49:43.539637202 +0000 UTC m=+1410.589428533" watchObservedRunningTime="2025-11-28 15:49:43.542975623 +0000 UTC m=+1410.592766934" Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.128688 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.129161 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.129554 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.129583 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.130546 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.134935 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.136968 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.137001 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.501737 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.533900 4805 generic.go:334] "Generic (PLEG): container finished" podID="bd26b960-0994-4bb3-aef3-035519e32420" containerID="59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a" exitCode=0 Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.534605 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.535105 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerDied","Data":"59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a"} Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.535142 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd26b960-0994-4bb3-aef3-035519e32420","Type":"ContainerDied","Data":"dcf4311d23e4526b4339c77ac45181acc6ef94781438d36bcef55e121805eb2c"} Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.535164 4805 scope.go:117] "RemoveContainer" containerID="6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.563835 4805 scope.go:117] "RemoveContainer" containerID="e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.585502 4805 scope.go:117] "RemoveContainer" containerID="59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.605475 4805 scope.go:117] "RemoveContainer" containerID="0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.637552 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-sg-core-conf-yaml\") pod \"bd26b960-0994-4bb3-aef3-035519e32420\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.637605 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-scripts\") pod \"bd26b960-0994-4bb3-aef3-035519e32420\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.637662 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-run-httpd\") pod \"bd26b960-0994-4bb3-aef3-035519e32420\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.637698 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7clt7\" (UniqueName: \"kubernetes.io/projected/bd26b960-0994-4bb3-aef3-035519e32420-kube-api-access-7clt7\") pod \"bd26b960-0994-4bb3-aef3-035519e32420\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.637730 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-combined-ca-bundle\") pod \"bd26b960-0994-4bb3-aef3-035519e32420\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.637762 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-log-httpd\") pod \"bd26b960-0994-4bb3-aef3-035519e32420\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.637843 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-ceilometer-tls-certs\") pod \"bd26b960-0994-4bb3-aef3-035519e32420\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.637876 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-config-data\") pod \"bd26b960-0994-4bb3-aef3-035519e32420\" (UID: \"bd26b960-0994-4bb3-aef3-035519e32420\") " Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.645669 4805 scope.go:117] "RemoveContainer" containerID="6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8" Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.649517 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8\": container with ID starting with 6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8 not found: ID does not exist" containerID="6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.649572 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8"} err="failed to get container status \"6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8\": rpc error: code = NotFound desc = could not find container \"6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8\": container with ID starting with 6ddaec5071f8cab9deebe084eb7377ebf16c8cd2af6fe9b4dc2f19fc0def39a8 not found: ID does not exist" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.649596 4805 scope.go:117] "RemoveContainer" containerID="e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.649675 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bd26b960-0994-4bb3-aef3-035519e32420" (UID: "bd26b960-0994-4bb3-aef3-035519e32420"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.649936 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bd26b960-0994-4bb3-aef3-035519e32420" (UID: "bd26b960-0994-4bb3-aef3-035519e32420"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.650266 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900\": container with ID starting with e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900 not found: ID does not exist" containerID="e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.650300 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900"} err="failed to get container status \"e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900\": rpc error: code = NotFound desc = could not find container \"e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900\": container with ID starting with e16156d98c1faaae35f7c66c1a61e358fbd640fb068a83b8390c2814fb487900 not found: ID does not exist" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.650324 4805 scope.go:117] "RemoveContainer" containerID="59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a" Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.650595 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a\": container with ID starting with 59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a not found: ID does not exist" containerID="59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.650618 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a"} err="failed to get container status \"59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a\": rpc error: code = NotFound desc = could not find container \"59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a\": container with ID starting with 59fb72b62e2fcd2a4a585e3747254011c5542ba1b4d2340a330f7cddb519ea1a not found: ID does not exist" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.650629 4805 scope.go:117] "RemoveContainer" containerID="0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c" Nov 28 15:49:44 crc kubenswrapper[4805]: E1128 15:49:44.651060 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c\": container with ID starting with 0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c not found: ID does not exist" containerID="0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.651082 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c"} err="failed to get container status \"0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c\": rpc error: code = NotFound desc = could not find container \"0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c\": container with ID starting with 0ddc97dd139c8da5873f5968259c673a45bd6b69778dc03f183230ab0663525c not found: ID does not exist" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.653123 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd26b960-0994-4bb3-aef3-035519e32420-kube-api-access-7clt7" (OuterVolumeSpecName: "kube-api-access-7clt7") pod "bd26b960-0994-4bb3-aef3-035519e32420" (UID: "bd26b960-0994-4bb3-aef3-035519e32420"). InnerVolumeSpecName "kube-api-access-7clt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.665568 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-scripts" (OuterVolumeSpecName: "scripts") pod "bd26b960-0994-4bb3-aef3-035519e32420" (UID: "bd26b960-0994-4bb3-aef3-035519e32420"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.696796 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bd26b960-0994-4bb3-aef3-035519e32420" (UID: "bd26b960-0994-4bb3-aef3-035519e32420"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.706482 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd26b960-0994-4bb3-aef3-035519e32420" (UID: "bd26b960-0994-4bb3-aef3-035519e32420"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.715706 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bd26b960-0994-4bb3-aef3-035519e32420" (UID: "bd26b960-0994-4bb3-aef3-035519e32420"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.744400 4805 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.744433 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.744444 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.744454 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.744464 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7clt7\" (UniqueName: \"kubernetes.io/projected/bd26b960-0994-4bb3-aef3-035519e32420-kube-api-access-7clt7\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.744478 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.744489 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd26b960-0994-4bb3-aef3-035519e32420-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.756801 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-config-data" (OuterVolumeSpecName: "config-data") pod "bd26b960-0994-4bb3-aef3-035519e32420" (UID: "bd26b960-0994-4bb3-aef3-035519e32420"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.845689 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd26b960-0994-4bb3-aef3-035519e32420-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.867352 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:49:44 crc kubenswrapper[4805]: I1128 15:49:44.873149 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:49:45 crc kubenswrapper[4805]: I1128 15:49:45.212526 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd26b960-0994-4bb3-aef3-035519e32420" path="/var/lib/kubelet/pods/bd26b960-0994-4bb3-aef3-035519e32420/volumes" Nov 28 15:49:45 crc kubenswrapper[4805]: I1128 15:49:45.894394 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-56786c7f4d-vg252" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.204:9311/healthcheck\": dial tcp 10.217.0.204:9311: i/o timeout" Nov 28 15:49:46 crc kubenswrapper[4805]: E1128 15:49:46.773932 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:49:46 crc kubenswrapper[4805]: E1128 15:49:46.774017 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts podName:2849f80f-ed31-4c7f-8f65-1132aa35b6e3 nodeName:}" failed. No retries permitted until 2025-11-28 15:49:54.77400214 +0000 UTC m=+1421.823793451 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts") pod "barbican8232-account-delete-52zhn" (UID: "2849f80f-ed31-4c7f-8f65-1132aa35b6e3") : configmap "openstack-scripts" not found Nov 28 15:49:48 crc kubenswrapper[4805]: E1128 15:49:48.210758 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:48 crc kubenswrapper[4805]: E1128 15:49:48.212004 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data podName:fb122aae-0e09-46b2-926c-037d25e79477 nodeName:}" failed. No retries permitted until 2025-11-28 15:50:04.21198432 +0000 UTC m=+1431.261775631 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb122aae-0e09-46b2-926c-037d25e79477") : configmap "rabbitmq-cell1-config-data" not found Nov 28 15:49:48 crc kubenswrapper[4805]: I1128 15:49:48.573682 4805 generic.go:334] "Generic (PLEG): container finished" podID="8724692d-29b7-46c8-b0dd-802e570050a7" containerID="7569e16b1b6d818541ba1419b8d0036700d58b74a89a37e7bc67e8d06bfaeb15" exitCode=0 Nov 28 15:49:48 crc kubenswrapper[4805]: I1128 15:49:48.573762 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdfcdc49-v4hgd" event={"ID":"8724692d-29b7-46c8-b0dd-802e570050a7","Type":"ContainerDied","Data":"7569e16b1b6d818541ba1419b8d0036700d58b74a89a37e7bc67e8d06bfaeb15"} Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.011124 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.115736 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.116291 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.116618 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.116684 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.118669 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.120239 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.121406 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.121449 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.124261 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-public-tls-certs\") pod \"8724692d-29b7-46c8-b0dd-802e570050a7\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.124323 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-combined-ca-bundle\") pod \"8724692d-29b7-46c8-b0dd-802e570050a7\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.124384 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-internal-tls-certs\") pod \"8724692d-29b7-46c8-b0dd-802e570050a7\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.124424 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cngjl\" (UniqueName: \"kubernetes.io/projected/8724692d-29b7-46c8-b0dd-802e570050a7-kube-api-access-cngjl\") pod \"8724692d-29b7-46c8-b0dd-802e570050a7\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.124450 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-httpd-config\") pod \"8724692d-29b7-46c8-b0dd-802e570050a7\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.124490 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-config\") pod \"8724692d-29b7-46c8-b0dd-802e570050a7\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.124522 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-ovndb-tls-certs\") pod \"8724692d-29b7-46c8-b0dd-802e570050a7\" (UID: \"8724692d-29b7-46c8-b0dd-802e570050a7\") " Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.130007 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "8724692d-29b7-46c8-b0dd-802e570050a7" (UID: "8724692d-29b7-46c8-b0dd-802e570050a7"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.131858 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8724692d-29b7-46c8-b0dd-802e570050a7-kube-api-access-cngjl" (OuterVolumeSpecName: "kube-api-access-cngjl") pod "8724692d-29b7-46c8-b0dd-802e570050a7" (UID: "8724692d-29b7-46c8-b0dd-802e570050a7"). InnerVolumeSpecName "kube-api-access-cngjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.169579 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8724692d-29b7-46c8-b0dd-802e570050a7" (UID: "8724692d-29b7-46c8-b0dd-802e570050a7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.171907 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8724692d-29b7-46c8-b0dd-802e570050a7" (UID: "8724692d-29b7-46c8-b0dd-802e570050a7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.172817 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8724692d-29b7-46c8-b0dd-802e570050a7" (UID: "8724692d-29b7-46c8-b0dd-802e570050a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.173982 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-config" (OuterVolumeSpecName: "config") pod "8724692d-29b7-46c8-b0dd-802e570050a7" (UID: "8724692d-29b7-46c8-b0dd-802e570050a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.191689 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "8724692d-29b7-46c8-b0dd-802e570050a7" (UID: "8724692d-29b7-46c8-b0dd-802e570050a7"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.225838 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.225869 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.225881 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.225895 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cngjl\" (UniqueName: \"kubernetes.io/projected/8724692d-29b7-46c8-b0dd-802e570050a7-kube-api-access-cngjl\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.225908 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.225919 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.225929 4805 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8724692d-29b7-46c8-b0dd-802e570050a7-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.583821 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdfcdc49-v4hgd" event={"ID":"8724692d-29b7-46c8-b0dd-802e570050a7","Type":"ContainerDied","Data":"6031a39a57464feb543c3973c7de58f3b2aac821b73b878c3d11850ed778203c"} Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.583873 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bdfcdc49-v4hgd" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.584253 4805 scope.go:117] "RemoveContainer" containerID="c253b5bf7e841bb27ed1346fc3c401ce938ec4f3dce125f741a8913badce2e99" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.604601 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6bdfcdc49-v4hgd"] Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.610274 4805 scope.go:117] "RemoveContainer" containerID="7569e16b1b6d818541ba1419b8d0036700d58b74a89a37e7bc67e8d06bfaeb15" Nov 28 15:49:49 crc kubenswrapper[4805]: I1128 15:49:49.610790 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6bdfcdc49-v4hgd"] Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.936147 4805 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.936206 4805 secret.go:188] Couldn't get secret openstack/barbican-keystone-listener-config-data: secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.936259 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:50:05.936242587 +0000 UTC m=+1432.986033908 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-keystone-listener-config-data" not found Nov 28 15:49:49 crc kubenswrapper[4805]: E1128 15:49:49.936277 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data podName:bbc55e90-2c22-4329-b3cc-375a68e6fe26 nodeName:}" failed. No retries permitted until 2025-11-28 15:50:05.936268518 +0000 UTC m=+1432.986059839 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data") pod "barbican-keystone-listener-7bfdbcc6c8-2pxb2" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26") : secret "barbican-config-data" not found Nov 28 15:49:51 crc kubenswrapper[4805]: I1128 15:49:51.214083 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" path="/var/lib/kubelet/pods/8724692d-29b7-46c8-b0dd-802e570050a7/volumes" Nov 28 15:49:53 crc kubenswrapper[4805]: I1128 15:49:53.161889 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:53 crc kubenswrapper[4805]: I1128 15:49:53.162259 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:53 crc kubenswrapper[4805]: I1128 15:49:53.230615 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:53 crc kubenswrapper[4805]: I1128 15:49:53.681415 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:53 crc kubenswrapper[4805]: I1128 15:49:53.742478 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nv5zc"] Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.115171 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.115865 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.116530 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.116612 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.117513 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.119074 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.121558 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.121805 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.833959 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:49:54 crc kubenswrapper[4805]: E1128 15:49:54.834067 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts podName:2849f80f-ed31-4c7f-8f65-1132aa35b6e3 nodeName:}" failed. No retries permitted until 2025-11-28 15:50:10.834049211 +0000 UTC m=+1437.883840532 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts") pod "barbican8232-account-delete-52zhn" (UID: "2849f80f-ed31-4c7f-8f65-1132aa35b6e3") : configmap "openstack-scripts" not found Nov 28 15:49:55 crc kubenswrapper[4805]: I1128 15:49:55.645144 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nv5zc" podUID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerName="registry-server" containerID="cri-o://553af6e1e47c786d07950057537bac41ba4e9ef9d1b92cfd40080a325fe92c1e" gracePeriod=2 Nov 28 15:49:57 crc kubenswrapper[4805]: I1128 15:49:57.690950 4805 generic.go:334] "Generic (PLEG): container finished" podID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerID="553af6e1e47c786d07950057537bac41ba4e9ef9d1b92cfd40080a325fe92c1e" exitCode=0 Nov 28 15:49:57 crc kubenswrapper[4805]: I1128 15:49:57.691121 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv5zc" event={"ID":"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5","Type":"ContainerDied","Data":"553af6e1e47c786d07950057537bac41ba4e9ef9d1b92cfd40080a325fe92c1e"} Nov 28 15:49:57 crc kubenswrapper[4805]: I1128 15:49:57.878480 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:57 crc kubenswrapper[4805]: I1128 15:49:57.997901 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-catalog-content\") pod \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " Nov 28 15:49:57 crc kubenswrapper[4805]: I1128 15:49:57.997968 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq79p\" (UniqueName: \"kubernetes.io/projected/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-kube-api-access-cq79p\") pod \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " Nov 28 15:49:57 crc kubenswrapper[4805]: I1128 15:49:57.998001 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-utilities\") pod \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\" (UID: \"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5\") " Nov 28 15:49:57 crc kubenswrapper[4805]: I1128 15:49:57.999110 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-utilities" (OuterVolumeSpecName: "utilities") pod "796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" (UID: "796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.003474 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-kube-api-access-cq79p" (OuterVolumeSpecName: "kube-api-access-cq79p") pod "796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" (UID: "796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5"). InnerVolumeSpecName "kube-api-access-cq79p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.099382 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq79p\" (UniqueName: \"kubernetes.io/projected/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-kube-api-access-cq79p\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.099410 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.122593 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" (UID: "796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.201177 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.703872 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv5zc" event={"ID":"796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5","Type":"ContainerDied","Data":"8b9784db64b8a1f7b20be90265955171bb18713e7d11151b51e3d2713273a642"} Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.703972 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv5zc" Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.704311 4805 scope.go:117] "RemoveContainer" containerID="553af6e1e47c786d07950057537bac41ba4e9ef9d1b92cfd40080a325fe92c1e" Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.746470 4805 scope.go:117] "RemoveContainer" containerID="2769fb70dd932ad97b359e1ae451227e6a8973823113353f6d855fb56e16ed14" Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.749559 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nv5zc"] Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.756836 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nv5zc"] Nov 28 15:49:58 crc kubenswrapper[4805]: I1128 15:49:58.776430 4805 scope.go:117] "RemoveContainer" containerID="d1b2e011187c0efc47f70c8cd2cec54317c7da156f1da659b8207f714d6684f5" Nov 28 15:49:59 crc kubenswrapper[4805]: E1128 15:49:59.115792 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:59 crc kubenswrapper[4805]: E1128 15:49:59.116164 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:59 crc kubenswrapper[4805]: E1128 15:49:59.116708 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 28 15:49:59 crc kubenswrapper[4805]: E1128 15:49:59.116798 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" Nov 28 15:49:59 crc kubenswrapper[4805]: E1128 15:49:59.117636 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:59 crc kubenswrapper[4805]: E1128 15:49:59.119506 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:59 crc kubenswrapper[4805]: E1128 15:49:59.121730 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 28 15:49:59 crc kubenswrapper[4805]: E1128 15:49:59.121791 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vssvf" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" Nov 28 15:49:59 crc kubenswrapper[4805]: I1128 15:49:59.217309 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" path="/var/lib/kubelet/pods/796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5/volumes" Nov 28 15:50:02 crc kubenswrapper[4805]: I1128 15:50:02.759311 4805 generic.go:334] "Generic (PLEG): container finished" podID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerID="4f7f7afc6966bf70692a1ec631ffcd04f68c6d162a03e988818a686b37ca27b5" exitCode=137 Nov 28 15:50:02 crc kubenswrapper[4805]: I1128 15:50:02.759344 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54ff5706-18a9-4e71-8289-e3d9e5d9af23","Type":"ContainerDied","Data":"4f7f7afc6966bf70692a1ec631ffcd04f68c6d162a03e988818a686b37ca27b5"} Nov 28 15:50:02 crc kubenswrapper[4805]: I1128 15:50:02.765801 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vssvf_89d793e5-c2d1-4630-95ff-615cd30b5d04/ovs-vswitchd/0.log" Nov 28 15:50:02 crc kubenswrapper[4805]: I1128 15:50:02.766898 4805 generic.go:334] "Generic (PLEG): container finished" podID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" exitCode=137 Nov 28 15:50:02 crc kubenswrapper[4805]: I1128 15:50:02.766940 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vssvf" event={"ID":"89d793e5-c2d1-4630-95ff-615cd30b5d04","Type":"ContainerDied","Data":"e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7"} Nov 28 15:50:02 crc kubenswrapper[4805]: I1128 15:50:02.778458 4805 generic.go:334] "Generic (PLEG): container finished" podID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerID="7413e61bbe43f8d00d9ac4ea2485eca35952ef72195b1eb40171e6705175ffcd" exitCode=137 Nov 28 15:50:02 crc kubenswrapper[4805]: I1128 15:50:02.778511 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"7413e61bbe43f8d00d9ac4ea2485eca35952ef72195b1eb40171e6705175ffcd"} Nov 28 15:50:02 crc kubenswrapper[4805]: I1128 15:50:02.997085 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.069743 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vssvf_89d793e5-c2d1-4630-95ff-615cd30b5d04/ovs-vswitchd/0.log" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.070467 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.087838 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-log\") pod \"89d793e5-c2d1-4630-95ff-615cd30b5d04\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.087909 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-combined-ca-bundle\") pod \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.087968 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w8rd\" (UniqueName: \"kubernetes.io/projected/54ff5706-18a9-4e71-8289-e3d9e5d9af23-kube-api-access-7w8rd\") pod \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088012 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-lib\") pod \"89d793e5-c2d1-4630-95ff-615cd30b5d04\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088037 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data\") pod \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088088 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data-custom\") pod \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088128 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d793e5-c2d1-4630-95ff-615cd30b5d04-scripts\") pod \"89d793e5-c2d1-4630-95ff-615cd30b5d04\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088174 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-run\") pod \"89d793e5-c2d1-4630-95ff-615cd30b5d04\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088247 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2v8k\" (UniqueName: \"kubernetes.io/projected/89d793e5-c2d1-4630-95ff-615cd30b5d04-kube-api-access-g2v8k\") pod \"89d793e5-c2d1-4630-95ff-615cd30b5d04\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088310 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-etc-ovs\") pod \"89d793e5-c2d1-4630-95ff-615cd30b5d04\" (UID: \"89d793e5-c2d1-4630-95ff-615cd30b5d04\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088351 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54ff5706-18a9-4e71-8289-e3d9e5d9af23-etc-machine-id\") pod \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088415 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-scripts\") pod \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\" (UID: \"54ff5706-18a9-4e71-8289-e3d9e5d9af23\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.088944 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-log" (OuterVolumeSpecName: "var-log") pod "89d793e5-c2d1-4630-95ff-615cd30b5d04" (UID: "89d793e5-c2d1-4630-95ff-615cd30b5d04"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.089559 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "89d793e5-c2d1-4630-95ff-615cd30b5d04" (UID: "89d793e5-c2d1-4630-95ff-615cd30b5d04"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.089615 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-run" (OuterVolumeSpecName: "var-run") pod "89d793e5-c2d1-4630-95ff-615cd30b5d04" (UID: "89d793e5-c2d1-4630-95ff-615cd30b5d04"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.089709 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54ff5706-18a9-4e71-8289-e3d9e5d9af23-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "54ff5706-18a9-4e71-8289-e3d9e5d9af23" (UID: "54ff5706-18a9-4e71-8289-e3d9e5d9af23"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.089782 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-lib" (OuterVolumeSpecName: "var-lib") pod "89d793e5-c2d1-4630-95ff-615cd30b5d04" (UID: "89d793e5-c2d1-4630-95ff-615cd30b5d04"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.091001 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89d793e5-c2d1-4630-95ff-615cd30b5d04-scripts" (OuterVolumeSpecName: "scripts") pod "89d793e5-c2d1-4630-95ff-615cd30b5d04" (UID: "89d793e5-c2d1-4630-95ff-615cd30b5d04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.095383 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "54ff5706-18a9-4e71-8289-e3d9e5d9af23" (UID: "54ff5706-18a9-4e71-8289-e3d9e5d9af23"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.095380 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89d793e5-c2d1-4630-95ff-615cd30b5d04-kube-api-access-g2v8k" (OuterVolumeSpecName: "kube-api-access-g2v8k") pod "89d793e5-c2d1-4630-95ff-615cd30b5d04" (UID: "89d793e5-c2d1-4630-95ff-615cd30b5d04"). InnerVolumeSpecName "kube-api-access-g2v8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.095451 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54ff5706-18a9-4e71-8289-e3d9e5d9af23-kube-api-access-7w8rd" (OuterVolumeSpecName: "kube-api-access-7w8rd") pod "54ff5706-18a9-4e71-8289-e3d9e5d9af23" (UID: "54ff5706-18a9-4e71-8289-e3d9e5d9af23"). InnerVolumeSpecName "kube-api-access-7w8rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.097978 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-scripts" (OuterVolumeSpecName: "scripts") pod "54ff5706-18a9-4e71-8289-e3d9e5d9af23" (UID: "54ff5706-18a9-4e71-8289-e3d9e5d9af23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.113728 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.138707 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54ff5706-18a9-4e71-8289-e3d9e5d9af23" (UID: "54ff5706-18a9-4e71-8289-e3d9e5d9af23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.189932 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8scq\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-kube-api-access-j8scq\") pod \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190004 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-cache\") pod \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190041 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") pod \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190085 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-lock\") pod \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190128 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\" (UID: \"f3d69b1a-cf80-424f-ac0c-7815e7ba0375\") " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190416 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w8rd\" (UniqueName: \"kubernetes.io/projected/54ff5706-18a9-4e71-8289-e3d9e5d9af23-kube-api-access-7w8rd\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190433 4805 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-lib\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190442 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190450 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d793e5-c2d1-4630-95ff-615cd30b5d04-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190458 4805 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190466 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2v8k\" (UniqueName: \"kubernetes.io/projected/89d793e5-c2d1-4630-95ff-615cd30b5d04-kube-api-access-g2v8k\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190475 4805 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-etc-ovs\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190483 4805 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54ff5706-18a9-4e71-8289-e3d9e5d9af23-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190490 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190501 4805 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89d793e5-c2d1-4630-95ff-615cd30b5d04-var-log\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.190509 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.191469 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-cache" (OuterVolumeSpecName: "cache") pod "f3d69b1a-cf80-424f-ac0c-7815e7ba0375" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.191959 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-lock" (OuterVolumeSpecName: "lock") pod "f3d69b1a-cf80-424f-ac0c-7815e7ba0375" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.193705 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f3d69b1a-cf80-424f-ac0c-7815e7ba0375" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.193798 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-kube-api-access-j8scq" (OuterVolumeSpecName: "kube-api-access-j8scq") pod "f3d69b1a-cf80-424f-ac0c-7815e7ba0375" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375"). InnerVolumeSpecName "kube-api-access-j8scq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.194043 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "swift") pod "f3d69b1a-cf80-424f-ac0c-7815e7ba0375" (UID: "f3d69b1a-cf80-424f-ac0c-7815e7ba0375"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.202049 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data" (OuterVolumeSpecName: "config-data") pod "54ff5706-18a9-4e71-8289-e3d9e5d9af23" (UID: "54ff5706-18a9-4e71-8289-e3d9e5d9af23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.292495 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8scq\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-kube-api-access-j8scq\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.292725 4805 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-cache\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.292823 4805 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.293055 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ff5706-18a9-4e71-8289-e3d9e5d9af23-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.293134 4805 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f3d69b1a-cf80-424f-ac0c-7815e7ba0375-lock\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.293217 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.308531 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.395658 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.793307 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.793328 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54ff5706-18a9-4e71-8289-e3d9e5d9af23","Type":"ContainerDied","Data":"7f867b3d73f37c5c0b2aaf08a77575ff2b6f0f5a9243efe6cda11c61e722f228"} Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.793508 4805 scope.go:117] "RemoveContainer" containerID="8007f0c48cb665204284cbdc1681cef5de09b06e0e0fbddc280c197c6fe89921" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.798404 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vssvf_89d793e5-c2d1-4630-95ff-615cd30b5d04/ovs-vswitchd/0.log" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.799590 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vssvf" event={"ID":"89d793e5-c2d1-4630-95ff-615cd30b5d04","Type":"ContainerDied","Data":"42e8b135ecdd2e74471d5c3cd1f1a2acbffee0ec25633ec02f3b3344957ef357"} Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.799782 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vssvf" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.816179 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.823850 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f3d69b1a-cf80-424f-ac0c-7815e7ba0375","Type":"ContainerDied","Data":"69fff0ee7eff65d669d3ec50922824c17777dd709a0a3ea5ac2e57087bb8a133"} Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.823999 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.827479 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.848030 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-vssvf"] Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.850992 4805 scope.go:117] "RemoveContainer" containerID="4f7f7afc6966bf70692a1ec631ffcd04f68c6d162a03e988818a686b37ca27b5" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.859257 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-vssvf"] Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.865632 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.871993 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.877884 4805 scope.go:117] "RemoveContainer" containerID="e74f6fa23a4141b9986355126441871718878e802820620f70c4d515b450f5f7" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.901849 4805 scope.go:117] "RemoveContainer" containerID="2873ae37e5c9b471b0c7188c139ea08e8742bea1d1177d42f0f9803360b8dce0" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.920376 4805 scope.go:117] "RemoveContainer" containerID="3752e4507479f7ee571985787e2818df7e3011388484910f0422a84a2d442fc5" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.943468 4805 scope.go:117] "RemoveContainer" containerID="7413e61bbe43f8d00d9ac4ea2485eca35952ef72195b1eb40171e6705175ffcd" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.963459 4805 scope.go:117] "RemoveContainer" containerID="841feae0d3f809ed9e79dff8941de56c70aafb0de8f991cd7453d48a9c67cc4c" Nov 28 15:50:03 crc kubenswrapper[4805]: I1128 15:50:03.988680 4805 scope.go:117] "RemoveContainer" containerID="7bc27ced72f3c03367ba7ebd0e3bbe80199cea1ba0a0d9701c61d76ee2259b8b" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.013698 4805 scope.go:117] "RemoveContainer" containerID="68a336481b5b62ed5a8243f246583d3e81d48c711638463168818d7ae197b2fe" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.034701 4805 scope.go:117] "RemoveContainer" containerID="4511ff31bb1feea8dab2de6f562fa4e27fc7ba92977a8fbc6bf26f4714ba0be0" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.056719 4805 scope.go:117] "RemoveContainer" containerID="7968f4c1d7d9d723379b1a1881a1e60cc6af6ce686c094d24597209d9d9c45cd" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.077958 4805 scope.go:117] "RemoveContainer" containerID="a92aa0f7330344e2f61dbf42e3f768648b41c7554cad6173713ec787c88993fc" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.094550 4805 scope.go:117] "RemoveContainer" containerID="fbaa7d95c8f6e7a6d621aa209aea65439fd1c81e062456fd47f282542a39aea3" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.113458 4805 scope.go:117] "RemoveContainer" containerID="7c0cf000e71a5c9d2c9ce21d71ecb228ec004603af520ea9f4b7232773601e18" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.131515 4805 scope.go:117] "RemoveContainer" containerID="10fe5f001457130421ee90b15666db980a7890c30af98e597a996d50a50c4e9a" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.148885 4805 scope.go:117] "RemoveContainer" containerID="831f948c09940fb083fcf963fed4d7066d04ebc8f2adc914bc711740e0fcdf49" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.165075 4805 scope.go:117] "RemoveContainer" containerID="663742bde9628596b637c938ce19b2d482d18491b8629d57987967e1b1b24891" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.187120 4805 scope.go:117] "RemoveContainer" containerID="d259dddfea43ce8231ff4ea937ce34f8e372f66bcd4c5effa7aec7ce0ffa8390" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.206225 4805 scope.go:117] "RemoveContainer" containerID="3ead42c6ea605c5d27888381393dc51bf09a9814de73cd305dbca46329a8a17d" Nov 28 15:50:04 crc kubenswrapper[4805]: I1128 15:50:04.236872 4805 scope.go:117] "RemoveContainer" containerID="cbabf490d47a58ab2c8c9f6eacb4cf048920fd4314a573b001947462a0de16b4" Nov 28 15:50:04 crc kubenswrapper[4805]: E1128 15:50:04.309429 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 15:50:04 crc kubenswrapper[4805]: E1128 15:50:04.309550 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data podName:fb122aae-0e09-46b2-926c-037d25e79477 nodeName:}" failed. No retries permitted until 2025-11-28 15:50:36.309494884 +0000 UTC m=+1463.359286195 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb122aae-0e09-46b2-926c-037d25e79477") : configmap "rabbitmq-cell1-config-data" not found Nov 28 15:50:05 crc kubenswrapper[4805]: I1128 15:50:05.220099 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" path="/var/lib/kubelet/pods/54ff5706-18a9-4e71-8289-e3d9e5d9af23/volumes" Nov 28 15:50:05 crc kubenswrapper[4805]: I1128 15:50:05.222066 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" path="/var/lib/kubelet/pods/89d793e5-c2d1-4630-95ff-615cd30b5d04/volumes" Nov 28 15:50:05 crc kubenswrapper[4805]: I1128 15:50:05.223715 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" path="/var/lib/kubelet/pods/f3d69b1a-cf80-424f-ac0c-7815e7ba0375/volumes" Nov 28 15:50:05 crc kubenswrapper[4805]: I1128 15:50:05.848683 4805 generic.go:334] "Generic (PLEG): container finished" podID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerID="5ac62896faa25edf067e9eba6f68e421a1b233d5e85b4ee414fe3961dc9ac78d" exitCode=137 Nov 28 15:50:05 crc kubenswrapper[4805]: I1128 15:50:05.848849 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" event={"ID":"bbc55e90-2c22-4329-b3cc-375a68e6fe26","Type":"ContainerDied","Data":"5ac62896faa25edf067e9eba6f68e421a1b233d5e85b4ee414fe3961dc9ac78d"} Nov 28 15:50:05 crc kubenswrapper[4805]: I1128 15:50:05.930107 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.035396 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom\") pod \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.035532 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc55e90-2c22-4329-b3cc-375a68e6fe26-logs\") pod \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.036006 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbc55e90-2c22-4329-b3cc-375a68e6fe26-logs" (OuterVolumeSpecName: "logs") pod "bbc55e90-2c22-4329-b3cc-375a68e6fe26" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.036100 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-combined-ca-bundle\") pod \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.036158 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hs4m\" (UniqueName: \"kubernetes.io/projected/bbc55e90-2c22-4329-b3cc-375a68e6fe26-kube-api-access-7hs4m\") pod \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.036194 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data\") pod \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\" (UID: \"bbc55e90-2c22-4329-b3cc-375a68e6fe26\") " Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.036513 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc55e90-2c22-4329-b3cc-375a68e6fe26-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.040493 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbc55e90-2c22-4329-b3cc-375a68e6fe26-kube-api-access-7hs4m" (OuterVolumeSpecName: "kube-api-access-7hs4m") pod "bbc55e90-2c22-4329-b3cc-375a68e6fe26" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26"). InnerVolumeSpecName "kube-api-access-7hs4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.040540 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bbc55e90-2c22-4329-b3cc-375a68e6fe26" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.060131 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbc55e90-2c22-4329-b3cc-375a68e6fe26" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.077092 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data" (OuterVolumeSpecName: "config-data") pod "bbc55e90-2c22-4329-b3cc-375a68e6fe26" (UID: "bbc55e90-2c22-4329-b3cc-375a68e6fe26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.138268 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.138322 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hs4m\" (UniqueName: \"kubernetes.io/projected/bbc55e90-2c22-4329-b3cc-375a68e6fe26-kube-api-access-7hs4m\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.138344 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.138386 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbc55e90-2c22-4329-b3cc-375a68e6fe26-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.860872 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" event={"ID":"bbc55e90-2c22-4329-b3cc-375a68e6fe26","Type":"ContainerDied","Data":"11c1ce48e069b9a3f885a452ae8ee6f1913b07862dfbb3a191d47dc83ecebfac"} Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.860936 4805 scope.go:117] "RemoveContainer" containerID="5ac62896faa25edf067e9eba6f68e421a1b233d5e85b4ee414fe3961dc9ac78d" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.860953 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.901916 4805 scope.go:117] "RemoveContainer" containerID="589d702195e92daa9f00af6a62686aa77c68b1a505d4b127244e5ebcb17036cd" Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.903410 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2"] Nov 28 15:50:06 crc kubenswrapper[4805]: I1128 15:50:06.911127 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7bfdbcc6c8-2pxb2"] Nov 28 15:50:07 crc kubenswrapper[4805]: I1128 15:50:07.218680 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" path="/var/lib/kubelet/pods/bbc55e90-2c22-4329-b3cc-375a68e6fe26/volumes" Nov 28 15:50:07 crc kubenswrapper[4805]: I1128 15:50:07.714233 4805 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pode50653a2-4881-42fd-bc8a-35e5f8e2f48d"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pode50653a2-4881-42fd-bc8a-35e5f8e2f48d] : Timed out while waiting for systemd to remove kubepods-besteffort-pode50653a2_4881_42fd_bc8a_35e5f8e2f48d.slice" Nov 28 15:50:07 crc kubenswrapper[4805]: E1128 15:50:07.714288 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pode50653a2-4881-42fd-bc8a-35e5f8e2f48d] : unable to destroy cgroup paths for cgroup [kubepods besteffort pode50653a2-4881-42fd-bc8a-35e5f8e2f48d] : Timed out while waiting for systemd to remove kubepods-besteffort-pode50653a2_4881_42fd_bc8a_35e5f8e2f48d.slice" pod="openstack/nova-cell1-novncproxy-0" podUID="e50653a2-4881-42fd-bc8a-35e5f8e2f48d" Nov 28 15:50:07 crc kubenswrapper[4805]: I1128 15:50:07.731464 4805 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podbb8cd2a8-99cd-47d2-ba18-c34a9b86abbc"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podbb8cd2a8-99cd-47d2-ba18-c34a9b86abbc] : Timed out while waiting for systemd to remove kubepods-besteffort-podbb8cd2a8_99cd_47d2_ba18_c34a9b86abbc.slice" Nov 28 15:50:07 crc kubenswrapper[4805]: I1128 15:50:07.736587 4805 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podb36ae016-8ef7-4c1f-a902-9e6f33d9aaad"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podb36ae016-8ef7-4c1f-a902-9e6f33d9aaad] : Timed out while waiting for systemd to remove kubepods-besteffort-podb36ae016_8ef7_4c1f_a902_9e6f33d9aaad.slice" Nov 28 15:50:07 crc kubenswrapper[4805]: I1128 15:50:07.738808 4805 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pode50653a2-4881-42fd-bc8a-35e5f8e2f48d"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pode50653a2-4881-42fd-bc8a-35e5f8e2f48d] : Timed out while waiting for systemd to remove kubepods-besteffort-pode50653a2_4881_42fd_bc8a_35e5f8e2f48d.slice" Nov 28 15:50:07 crc kubenswrapper[4805]: I1128 15:50:07.872324 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 15:50:07 crc kubenswrapper[4805]: I1128 15:50:07.898655 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:50:07 crc kubenswrapper[4805]: I1128 15:50:07.904331 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.216856 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e50653a2-4881-42fd-bc8a-35e5f8e2f48d" path="/var/lib/kubelet/pods/e50653a2-4881-42fd-bc8a-35e5f8e2f48d/volumes" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.528908 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.607583 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data-custom\") pod \"95167e93-42b8-4f5f-b5a9-587b9b854f12\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.607719 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-combined-ca-bundle\") pod \"95167e93-42b8-4f5f-b5a9-587b9b854f12\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.607775 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data\") pod \"95167e93-42b8-4f5f-b5a9-587b9b854f12\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.607812 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95167e93-42b8-4f5f-b5a9-587b9b854f12-logs\") pod \"95167e93-42b8-4f5f-b5a9-587b9b854f12\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.607874 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf5wk\" (UniqueName: \"kubernetes.io/projected/95167e93-42b8-4f5f-b5a9-587b9b854f12-kube-api-access-nf5wk\") pod \"95167e93-42b8-4f5f-b5a9-587b9b854f12\" (UID: \"95167e93-42b8-4f5f-b5a9-587b9b854f12\") " Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.608872 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95167e93-42b8-4f5f-b5a9-587b9b854f12-logs" (OuterVolumeSpecName: "logs") pod "95167e93-42b8-4f5f-b5a9-587b9b854f12" (UID: "95167e93-42b8-4f5f-b5a9-587b9b854f12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.614863 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95167e93-42b8-4f5f-b5a9-587b9b854f12-kube-api-access-nf5wk" (OuterVolumeSpecName: "kube-api-access-nf5wk") pod "95167e93-42b8-4f5f-b5a9-587b9b854f12" (UID: "95167e93-42b8-4f5f-b5a9-587b9b854f12"). InnerVolumeSpecName "kube-api-access-nf5wk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.621239 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "95167e93-42b8-4f5f-b5a9-587b9b854f12" (UID: "95167e93-42b8-4f5f-b5a9-587b9b854f12"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.631996 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95167e93-42b8-4f5f-b5a9-587b9b854f12" (UID: "95167e93-42b8-4f5f-b5a9-587b9b854f12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.651420 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data" (OuterVolumeSpecName: "config-data") pod "95167e93-42b8-4f5f-b5a9-587b9b854f12" (UID: "95167e93-42b8-4f5f-b5a9-587b9b854f12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.709422 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.709466 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.709479 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95167e93-42b8-4f5f-b5a9-587b9b854f12-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.709491 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95167e93-42b8-4f5f-b5a9-587b9b854f12-logs\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.709504 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf5wk\" (UniqueName: \"kubernetes.io/projected/95167e93-42b8-4f5f-b5a9-587b9b854f12-kube-api-access-nf5wk\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.827131 4805 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podac086fdf-9a1a-40ab-ab2f-8a33dae28290"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podac086fdf-9a1a-40ab-ab2f-8a33dae28290] : Timed out while waiting for systemd to remove kubepods-besteffort-podac086fdf_9a1a_40ab_ab2f_8a33dae28290.slice" Nov 28 15:50:09 crc kubenswrapper[4805]: E1128 15:50:09.827376 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podac086fdf-9a1a-40ab-ab2f-8a33dae28290] : unable to destroy cgroup paths for cgroup [kubepods besteffort podac086fdf-9a1a-40ab-ab2f-8a33dae28290] : Timed out while waiting for systemd to remove kubepods-besteffort-podac086fdf_9a1a_40ab_ab2f_8a33dae28290.slice" pod="openstack/nova-scheduler-0" podUID="ac086fdf-9a1a-40ab-ab2f-8a33dae28290" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.895598 4805 generic.go:334] "Generic (PLEG): container finished" podID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerID="0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c" exitCode=137 Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.895707 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.895741 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7849b77bf-z2qp2" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.895732 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7849b77bf-z2qp2" event={"ID":"95167e93-42b8-4f5f-b5a9-587b9b854f12","Type":"ContainerDied","Data":"0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c"} Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.895848 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7849b77bf-z2qp2" event={"ID":"95167e93-42b8-4f5f-b5a9-587b9b854f12","Type":"ContainerDied","Data":"fb34bb50b63f1a2069a8960d2e187651a99a18f19f886a0e8973572b03b7950f"} Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.895883 4805 scope.go:117] "RemoveContainer" containerID="0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.927556 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.936282 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.941939 4805 scope.go:117] "RemoveContainer" containerID="3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.944498 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7849b77bf-z2qp2"] Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.952252 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-7849b77bf-z2qp2"] Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.965311 4805 scope.go:117] "RemoveContainer" containerID="0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c" Nov 28 15:50:09 crc kubenswrapper[4805]: E1128 15:50:09.965762 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c\": container with ID starting with 0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c not found: ID does not exist" containerID="0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.965797 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c"} err="failed to get container status \"0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c\": rpc error: code = NotFound desc = could not find container \"0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c\": container with ID starting with 0457cfc81e8fd3c39c1c10e3f4d9f67cdb6064cd8c8cc6c7dbb5515a875fa48c not found: ID does not exist" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.965819 4805 scope.go:117] "RemoveContainer" containerID="3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf" Nov 28 15:50:09 crc kubenswrapper[4805]: E1128 15:50:09.966267 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf\": container with ID starting with 3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf not found: ID does not exist" containerID="3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf" Nov 28 15:50:09 crc kubenswrapper[4805]: I1128 15:50:09.966287 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf"} err="failed to get container status \"3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf\": rpc error: code = NotFound desc = could not find container \"3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf\": container with ID starting with 3e048580b6006059516454f47625eea2c7d2f87ac1af86b43c9512efa28a89cf not found: ID does not exist" Nov 28 15:50:10 crc kubenswrapper[4805]: E1128 15:50:10.927066 4805 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 28 15:50:10 crc kubenswrapper[4805]: E1128 15:50:10.927147 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts podName:2849f80f-ed31-4c7f-8f65-1132aa35b6e3 nodeName:}" failed. No retries permitted until 2025-11-28 15:50:42.927130213 +0000 UTC m=+1469.976921524 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts") pod "barbican8232-account-delete-52zhn" (UID: "2849f80f-ed31-4c7f-8f65-1132aa35b6e3") : configmap "openstack-scripts" not found Nov 28 15:50:11 crc kubenswrapper[4805]: I1128 15:50:11.059862 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:50:11 crc kubenswrapper[4805]: I1128 15:50:11.059929 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:50:11 crc kubenswrapper[4805]: I1128 15:50:11.212954 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95167e93-42b8-4f5f-b5a9-587b9b854f12" path="/var/lib/kubelet/pods/95167e93-42b8-4f5f-b5a9-587b9b854f12/volumes" Nov 28 15:50:11 crc kubenswrapper[4805]: I1128 15:50:11.213754 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac086fdf-9a1a-40ab-ab2f-8a33dae28290" path="/var/lib/kubelet/pods/ac086fdf-9a1a-40ab-ab2f-8a33dae28290/volumes" Nov 28 15:50:11 crc kubenswrapper[4805]: I1128 15:50:11.924770 4805 generic.go:334] "Generic (PLEG): container finished" podID="2849f80f-ed31-4c7f-8f65-1132aa35b6e3" containerID="ac4df6496bd374f34589a6d93fe9c479a53303f520c345b7a742bf81cedf5118" exitCode=137 Nov 28 15:50:11 crc kubenswrapper[4805]: I1128 15:50:11.924863 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8232-account-delete-52zhn" event={"ID":"2849f80f-ed31-4c7f-8f65-1132aa35b6e3","Type":"ContainerDied","Data":"ac4df6496bd374f34589a6d93fe9c479a53303f520c345b7a742bf81cedf5118"} Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.171979 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.243652 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts\") pod \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\" (UID: \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\") " Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.243728 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnz54\" (UniqueName: \"kubernetes.io/projected/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-kube-api-access-gnz54\") pod \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\" (UID: \"2849f80f-ed31-4c7f-8f65-1132aa35b6e3\") " Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.244541 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2849f80f-ed31-4c7f-8f65-1132aa35b6e3" (UID: "2849f80f-ed31-4c7f-8f65-1132aa35b6e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.256618 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-kube-api-access-gnz54" (OuterVolumeSpecName: "kube-api-access-gnz54") pod "2849f80f-ed31-4c7f-8f65-1132aa35b6e3" (UID: "2849f80f-ed31-4c7f-8f65-1132aa35b6e3"). InnerVolumeSpecName "kube-api-access-gnz54". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.345558 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.345917 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnz54\" (UniqueName: \"kubernetes.io/projected/2849f80f-ed31-4c7f-8f65-1132aa35b6e3-kube-api-access-gnz54\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.935234 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8232-account-delete-52zhn" event={"ID":"2849f80f-ed31-4c7f-8f65-1132aa35b6e3","Type":"ContainerDied","Data":"a0811251e175347c191f76c64c954e29e72a8d39c458b0611310ffcaa4e72667"} Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.935266 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8232-account-delete-52zhn" Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.935289 4805 scope.go:117] "RemoveContainer" containerID="ac4df6496bd374f34589a6d93fe9c479a53303f520c345b7a742bf81cedf5118" Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.967014 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8232-account-delete-52zhn"] Nov 28 15:50:12 crc kubenswrapper[4805]: I1128 15:50:12.974131 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican8232-account-delete-52zhn"] Nov 28 15:50:13 crc kubenswrapper[4805]: I1128 15:50:13.221063 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2849f80f-ed31-4c7f-8f65-1132aa35b6e3" path="/var/lib/kubelet/pods/2849f80f-ed31-4c7f-8f65-1132aa35b6e3/volumes" Nov 28 15:50:34 crc kubenswrapper[4805]: I1128 15:50:34.044501 4805 scope.go:117] "RemoveContainer" containerID="8aefb106dad36ecb957abbc418ec13004a00166941cb8f687492113dfe27ca6f" Nov 28 15:50:34 crc kubenswrapper[4805]: I1128 15:50:34.081423 4805 scope.go:117] "RemoveContainer" containerID="834da8fff3dc97a83b4cef8f408cff6c49dd4c0dba0419975a5a61642d26191a" Nov 28 15:50:36 crc kubenswrapper[4805]: E1128 15:50:36.400019 4805 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 28 15:50:36 crc kubenswrapper[4805]: E1128 15:50:36.400440 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data podName:fb122aae-0e09-46b2-926c-037d25e79477 nodeName:}" failed. No retries permitted until 2025-11-28 15:51:40.400421245 +0000 UTC m=+1527.450212556 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb122aae-0e09-46b2-926c-037d25e79477") : configmap "rabbitmq-cell1-config-data" not found Nov 28 15:50:36 crc kubenswrapper[4805]: E1128 15:50:36.557652 4805 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 28 15:50:36 crc kubenswrapper[4805]: command '/bin/bash -c if [ ! -z "$(cat /etc/pod-info/skipPreStopChecks)" ]; then exit 0; fi; rabbitmq-upgrade await_online_quorum_plus_one -t 604800 && rabbitmq-upgrade await_online_synchronized_mirror -t 604800 || true && rabbitmq-upgrade drain -t 604800' exited with 69: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Most common reasons for this are: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 28 15:50:36 crc kubenswrapper[4805]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is not running Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: In addition to the diagnostics info below: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 28 15:50:36 crc kubenswrapper[4805]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 28 15:50:36 crc kubenswrapper[4805]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: DIAGNOSTICS Nov 28 15:50:36 crc kubenswrapper[4805]: =========== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 28 15:50:36 crc kubenswrapper[4805]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Current node details: Nov 28 15:50:36 crc kubenswrapper[4805]: * node name: 'rabbitmqcli-477-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 28 15:50:36 crc kubenswrapper[4805]: * effective user's home directory: /var/lib/rabbitmq Nov 28 15:50:36 crc kubenswrapper[4805]: * Erlang cookie hash: COi4P+wFWOsQHfJalkvHNw== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Most common reasons for this are: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 28 15:50:36 crc kubenswrapper[4805]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is not running Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: In addition to the diagnostics info below: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 28 15:50:36 crc kubenswrapper[4805]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 28 15:50:36 crc kubenswrapper[4805]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: DIAGNOSTICS Nov 28 15:50:36 crc kubenswrapper[4805]: =========== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 28 15:50:36 crc kubenswrapper[4805]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Current node details: Nov 28 15:50:36 crc kubenswrapper[4805]: * node name: 'rabbitmqcli-383-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 28 15:50:36 crc kubenswrapper[4805]: * effective user's home directory: /var/lib/rabbitmq Nov 28 15:50:36 crc kubenswrapper[4805]: * Erlang cookie hash: COi4P+wFWOsQHfJalkvHNw== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: > execCommand=["/bin/bash","-c","if [ ! -z \"$(cat /etc/pod-info/skipPreStopChecks)\" ]; then exit 0; fi; rabbitmq-upgrade await_online_quorum_plus_one -t 604800 \u0026\u0026 rabbitmq-upgrade await_online_synchronized_mirror -t 604800 || true \u0026\u0026 rabbitmq-upgrade drain -t 604800"] containerName="rabbitmq" pod="openstack/rabbitmq-cell1-server-0" message=< Nov 28 15:50:36 crc kubenswrapper[4805]: Will put node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack into maintenance mode. The node will no longer serve any client traffic! Nov 28 15:50:36 crc kubenswrapper[4805]: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Most common reasons for this are: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 28 15:50:36 crc kubenswrapper[4805]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is not running Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: In addition to the diagnostics info below: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 28 15:50:36 crc kubenswrapper[4805]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 28 15:50:36 crc kubenswrapper[4805]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: DIAGNOSTICS Nov 28 15:50:36 crc kubenswrapper[4805]: =========== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 28 15:50:36 crc kubenswrapper[4805]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Current node details: Nov 28 15:50:36 crc kubenswrapper[4805]: * node name: 'rabbitmqcli-477-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 28 15:50:36 crc kubenswrapper[4805]: * effective user's home directory: /var/lib/rabbitmq Nov 28 15:50:36 crc kubenswrapper[4805]: * Erlang cookie hash: COi4P+wFWOsQHfJalkvHNw== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Most common reasons for this are: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 28 15:50:36 crc kubenswrapper[4805]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is not running Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: In addition to the diagnostics info below: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 28 15:50:36 crc kubenswrapper[4805]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 28 15:50:36 crc kubenswrapper[4805]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: DIAGNOSTICS Nov 28 15:50:36 crc kubenswrapper[4805]: =========== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 28 15:50:36 crc kubenswrapper[4805]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Current node details: Nov 28 15:50:36 crc kubenswrapper[4805]: * node name: 'rabbitmqcli-383-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 28 15:50:36 crc kubenswrapper[4805]: * effective user's home directory: /var/lib/rabbitmq Nov 28 15:50:36 crc kubenswrapper[4805]: * Erlang cookie hash: COi4P+wFWOsQHfJalkvHNw== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: > Nov 28 15:50:36 crc kubenswrapper[4805]: E1128 15:50:36.557705 4805 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 28 15:50:36 crc kubenswrapper[4805]: command '/bin/bash -c if [ ! -z "$(cat /etc/pod-info/skipPreStopChecks)" ]; then exit 0; fi; rabbitmq-upgrade await_online_quorum_plus_one -t 604800 && rabbitmq-upgrade await_online_synchronized_mirror -t 604800 || true && rabbitmq-upgrade drain -t 604800' exited with 69: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Most common reasons for this are: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 28 15:50:36 crc kubenswrapper[4805]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is not running Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: In addition to the diagnostics info below: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 28 15:50:36 crc kubenswrapper[4805]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 28 15:50:36 crc kubenswrapper[4805]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: DIAGNOSTICS Nov 28 15:50:36 crc kubenswrapper[4805]: =========== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 28 15:50:36 crc kubenswrapper[4805]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Current node details: Nov 28 15:50:36 crc kubenswrapper[4805]: * node name: 'rabbitmqcli-477-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 28 15:50:36 crc kubenswrapper[4805]: * effective user's home directory: /var/lib/rabbitmq Nov 28 15:50:36 crc kubenswrapper[4805]: * Erlang cookie hash: COi4P+wFWOsQHfJalkvHNw== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Most common reasons for this are: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 28 15:50:36 crc kubenswrapper[4805]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 28 15:50:36 crc kubenswrapper[4805]: * Target node is not running Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: In addition to the diagnostics info below: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 28 15:50:36 crc kubenswrapper[4805]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 28 15:50:36 crc kubenswrapper[4805]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: DIAGNOSTICS Nov 28 15:50:36 crc kubenswrapper[4805]: =========== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 28 15:50:36 crc kubenswrapper[4805]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: Current node details: Nov 28 15:50:36 crc kubenswrapper[4805]: * node name: 'rabbitmqcli-383-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 28 15:50:36 crc kubenswrapper[4805]: * effective user's home directory: /var/lib/rabbitmq Nov 28 15:50:36 crc kubenswrapper[4805]: * Erlang cookie hash: COi4P+wFWOsQHfJalkvHNw== Nov 28 15:50:36 crc kubenswrapper[4805]: Nov 28 15:50:36 crc kubenswrapper[4805]: > pod="openstack/rabbitmq-cell1-server-0" podUID="fb122aae-0e09-46b2-926c-037d25e79477" containerName="rabbitmq" containerID="cri-o://8c7376fa9dd5297887afcd13bb656b98782355f020c1ad2e8dfa78a8d4ee511a" Nov 28 15:50:36 crc kubenswrapper[4805]: I1128 15:50:36.557750 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="fb122aae-0e09-46b2-926c-037d25e79477" containerName="rabbitmq" containerID="cri-o://8c7376fa9dd5297887afcd13bb656b98782355f020c1ad2e8dfa78a8d4ee511a" gracePeriod=604739 Nov 28 15:50:41 crc kubenswrapper[4805]: I1128 15:50:41.060079 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:50:41 crc kubenswrapper[4805]: I1128 15:50:41.060352 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.218045 4805 generic.go:334] "Generic (PLEG): container finished" podID="fb122aae-0e09-46b2-926c-037d25e79477" containerID="8c7376fa9dd5297887afcd13bb656b98782355f020c1ad2e8dfa78a8d4ee511a" exitCode=0 Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.218095 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb122aae-0e09-46b2-926c-037d25e79477","Type":"ContainerDied","Data":"8c7376fa9dd5297887afcd13bb656b98782355f020c1ad2e8dfa78a8d4ee511a"} Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.218762 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb122aae-0e09-46b2-926c-037d25e79477","Type":"ContainerDied","Data":"90c5973fc0ce456c98a3efe57e0b031cf0efbc7f1b4525d3b71703c53e2ace07"} Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.218787 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90c5973fc0ce456c98a3efe57e0b031cf0efbc7f1b4525d3b71703c53e2ace07" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.222381 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304388 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb122aae-0e09-46b2-926c-037d25e79477-pod-info\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304449 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb122aae-0e09-46b2-926c-037d25e79477-erlang-cookie-secret\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304476 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzjl9\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-kube-api-access-kzjl9\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304549 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-plugins\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304610 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304641 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-plugins-conf\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304665 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-tls\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304734 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-erlang-cookie\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304752 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-confd\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.304939 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.305030 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-server-conf\") pod \"fb122aae-0e09-46b2-926c-037d25e79477\" (UID: \"fb122aae-0e09-46b2-926c-037d25e79477\") " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.306873 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.306901 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.307709 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.312473 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/fb122aae-0e09-46b2-926c-037d25e79477-pod-info" (OuterVolumeSpecName: "pod-info") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.312812 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb122aae-0e09-46b2-926c-037d25e79477-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.312851 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.313430 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.324766 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-kube-api-access-kzjl9" (OuterVolumeSpecName: "kube-api-access-kzjl9") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "kube-api-access-kzjl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.325983 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data" (OuterVolumeSpecName: "config-data") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.352598 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-server-conf" (OuterVolumeSpecName: "server-conf") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.387023 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "fb122aae-0e09-46b2-926c-037d25e79477" (UID: "fb122aae-0e09-46b2-926c-037d25e79477"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.406958 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407007 4805 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407019 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407031 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407043 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407054 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407062 4805 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb122aae-0e09-46b2-926c-037d25e79477-server-conf\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407071 4805 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb122aae-0e09-46b2-926c-037d25e79477-pod-info\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407079 4805 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb122aae-0e09-46b2-926c-037d25e79477-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407087 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzjl9\" (UniqueName: \"kubernetes.io/projected/fb122aae-0e09-46b2-926c-037d25e79477-kube-api-access-kzjl9\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.407095 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb122aae-0e09-46b2-926c-037d25e79477-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.420826 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 28 15:50:43 crc kubenswrapper[4805]: I1128 15:50:43.508838 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:50:44 crc kubenswrapper[4805]: I1128 15:50:44.232836 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 15:50:44 crc kubenswrapper[4805]: I1128 15:50:44.279335 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 15:50:44 crc kubenswrapper[4805]: I1128 15:50:44.289080 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 15:50:45 crc kubenswrapper[4805]: I1128 15:50:45.220485 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb122aae-0e09-46b2-926c-037d25e79477" path="/var/lib/kubelet/pods/fb122aae-0e09-46b2-926c-037d25e79477/volumes" Nov 28 15:51:11 crc kubenswrapper[4805]: I1128 15:51:11.060807 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:51:11 crc kubenswrapper[4805]: I1128 15:51:11.061423 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:51:11 crc kubenswrapper[4805]: I1128 15:51:11.061482 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:51:11 crc kubenswrapper[4805]: I1128 15:51:11.062138 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed"} pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:51:11 crc kubenswrapper[4805]: I1128 15:51:11.062188 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" containerID="cri-o://c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" gracePeriod=600 Nov 28 15:51:11 crc kubenswrapper[4805]: E1128 15:51:11.228896 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:51:11 crc kubenswrapper[4805]: I1128 15:51:11.504225 4805 generic.go:334] "Generic (PLEG): container finished" podID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" exitCode=0 Nov 28 15:51:11 crc kubenswrapper[4805]: I1128 15:51:11.504264 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerDied","Data":"c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed"} Nov 28 15:51:11 crc kubenswrapper[4805]: I1128 15:51:11.504293 4805 scope.go:117] "RemoveContainer" containerID="f2d2222cce3bf61f8537971c6c7eb69d6ee1501220244e7d4ba5697a12f1d2d5" Nov 28 15:51:11 crc kubenswrapper[4805]: I1128 15:51:11.504881 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:51:11 crc kubenswrapper[4805]: E1128 15:51:11.505162 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:51:24 crc kubenswrapper[4805]: I1128 15:51:24.204084 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:51:24 crc kubenswrapper[4805]: E1128 15:51:24.204934 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.666227 4805 scope.go:117] "RemoveContainer" containerID="4e4402e0fc9b295b789402eb980861bc790e80c1e7f03bfe24e34dbc5607e0b8" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.719633 4805 scope.go:117] "RemoveContainer" containerID="6ac18cc12cbd5e678d45384978eaba73c92c3dd0e8f7dfc622bed25ab557e718" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.745556 4805 scope.go:117] "RemoveContainer" containerID="01b2f5f82ace63d29f30997e3ed6d0a996047e0556752afaa1bc33b25beb49ca" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.779545 4805 scope.go:117] "RemoveContainer" containerID="1d33f1bb0bf0611e99c60c4a3164ece57f7449f7387ff2182d439cdcab248b27" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.795062 4805 scope.go:117] "RemoveContainer" containerID="5dec4cf0f7bf4573bbaa714bf92c31c74fecb1cf0fab84de687c65be40e99c1f" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.822951 4805 scope.go:117] "RemoveContainer" containerID="32d65daedeab95d64fa9e1a47825872f67c3ce768660a116428eaf18af16ec3f" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.850791 4805 scope.go:117] "RemoveContainer" containerID="1ccc26baa6ad7cbc405abb6d95711fcbf099c77d7ab2dbe23883e5b60b452fc1" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.870479 4805 scope.go:117] "RemoveContainer" containerID="cc6c21d34e2ac7e3ae4c9f58a0944af49da732e4fdc04209652d7d7e9cd2a44e" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.892534 4805 scope.go:117] "RemoveContainer" containerID="14a51093a1b996f3d81f038eac81b07209514a1e4e613ae566cfbba1064f591f" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.920126 4805 scope.go:117] "RemoveContainer" containerID="165b9749994c0b329221bf6754fc0338f2accf4df65b8eaf405f46a8898dbdaf" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.940517 4805 scope.go:117] "RemoveContainer" containerID="4d4f71d0ee96e28e633cd614ac2ab22db848b839df7b6c625454acce90e318e8" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.955852 4805 scope.go:117] "RemoveContainer" containerID="ac30b0deb20776f06cef1ad68dfa668fd58449a1db45c10823b51c020d31976c" Nov 28 15:51:34 crc kubenswrapper[4805]: I1128 15:51:34.974811 4805 scope.go:117] "RemoveContainer" containerID="7f4ba1f787d29d980bc0c209ec06ceece827e2ac14219237aa842b0f992266a1" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.028277 4805 scope.go:117] "RemoveContainer" containerID="aa9513cfb6de7702a6b4bec85ab8f4e05e6b8e156f22d6759ee68c0788aa186f" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.044365 4805 scope.go:117] "RemoveContainer" containerID="3aee0668fd5c39ac3a3e4f335457cb6946f12c017da091b68fe6eda24c00f29b" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.092111 4805 scope.go:117] "RemoveContainer" containerID="0ebb89ebcccc94757bb175b0c950e5a995e346c32fea6d70aa71be39c4c62c8d" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.112375 4805 scope.go:117] "RemoveContainer" containerID="65d607a09a90f0f414b90b8d0bc2013bb6501ea2d7d313a6ba0262c0e54abf2d" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.138489 4805 scope.go:117] "RemoveContainer" containerID="8c7376fa9dd5297887afcd13bb656b98782355f020c1ad2e8dfa78a8d4ee511a" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.160541 4805 scope.go:117] "RemoveContainer" containerID="5728fc21b8a8118968ef55e7ad974f51510ac290bf4297215efb9f590e045ec0" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.180258 4805 scope.go:117] "RemoveContainer" containerID="25f906e361ec4998574fde2ff49ffc6709f4fa6fb833757df62d571312c0b805" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.199051 4805 scope.go:117] "RemoveContainer" containerID="ebdb3b51aa00ed05f41aa993d8ac77042a1ba60f7c7205c1ecaf17706bd60c87" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.204974 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:51:35 crc kubenswrapper[4805]: E1128 15:51:35.205404 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.221765 4805 scope.go:117] "RemoveContainer" containerID="9043d7fdde1ce41cbf522437ac97e3b9791a64f0d7098912354924104f8f16f6" Nov 28 15:51:35 crc kubenswrapper[4805]: I1128 15:51:35.246257 4805 scope.go:117] "RemoveContainer" containerID="7a66ca4e8862507987fdb88cdd304c61afe27c97c176cd5e1cfb6c13d29f96ad" Nov 28 15:51:50 crc kubenswrapper[4805]: I1128 15:51:50.204143 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:51:50 crc kubenswrapper[4805]: E1128 15:51:50.205152 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.644426 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bstbq"] Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.644976 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db4967f-554e-4336-913a-fcebe420cf0e" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.644989 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db4967f-554e-4336-913a-fcebe420cf0e" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645006 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645012 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645019 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="614a68ec-3129-413f-abb1-40a73ad9137e" containerName="keystone-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645028 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="614a68ec-3129-413f-abb1-40a73ad9137e" containerName="keystone-api" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645037 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerName="barbican-worker-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645043 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerName="barbican-worker-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645051 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-metadata" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645056 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-metadata" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645067 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-updater" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645073 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-updater" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645079 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f912081e-175b-410c-bbfa-daa3a71e2179" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645084 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f912081e-175b-410c-bbfa-daa3a71e2179" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645093 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2849f80f-ed31-4c7f-8f65-1132aa35b6e3" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645099 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2849f80f-ed31-4c7f-8f65-1132aa35b6e3" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645108 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerName="barbican-keystone-listener-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645114 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerName="barbican-keystone-listener-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645126 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645133 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645144 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerName="extract-utilities" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645149 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerName="extract-utilities" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645156 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" containerName="neutron-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645161 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" containerName="neutron-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645170 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" containerName="nova-cell0-conductor-conductor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645176 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" containerName="nova-cell0-conductor-conductor" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645186 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-updater" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645193 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-updater" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645202 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624edf72-de15-4026-812b-36d993917176" containerName="ovn-northd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645207 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="624edf72-de15-4026-812b-36d993917176" containerName="ovn-northd" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645219 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="ceilometer-central-agent" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645225 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="ceilometer-central-agent" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645232 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerName="barbican-worker-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645237 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerName="barbican-worker-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645246 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-replicator" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645251 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-replicator" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645259 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645266 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645274 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerName="setup-container" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645279 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerName="setup-container" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645286 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645292 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645302 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerName="placement-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645308 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerName="placement-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645319 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerName="registry-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645325 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerName="registry-server" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645335 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb122aae-0e09-46b2-926c-037d25e79477" containerName="setup-container" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645340 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb122aae-0e09-46b2-926c-037d25e79477" containerName="setup-container" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645348 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd6fbaf-5702-415e-8a1d-85d2c4792031" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645374 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd6fbaf-5702-415e-8a1d-85d2c4792031" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645389 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerName="glance-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645395 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerName="glance-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645407 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerName="glance-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645413 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerName="glance-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645420 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="ceilometer-notification-agent" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645425 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="ceilometer-notification-agent" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645437 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-auditor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645443 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-auditor" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645450 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645456 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645463 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf2e5369-a11b-4150-a291-4aeab9724d82" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645468 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf2e5369-a11b-4150-a291-4aeab9724d82" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645477 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerName="cinder-scheduler" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645483 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerName="cinder-scheduler" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645490 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d7a9256-533a-4985-9a62-9bbe76af9e54" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645495 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d7a9256-533a-4985-9a62-9bbe76af9e54" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645505 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" containerName="mysql-bootstrap" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645511 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" containerName="mysql-bootstrap" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645520 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-replicator" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645526 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-replicator" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645533 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452c347f-4ee3-46de-ba8f-c83300966f5d" containerName="mysql-bootstrap" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645538 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="452c347f-4ee3-46de-ba8f-c83300966f5d" containerName="mysql-bootstrap" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645546 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerName="glance-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645553 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerName="glance-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645562 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" containerName="galera" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645569 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" containerName="galera" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645582 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645590 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-server" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645599 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645606 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645616 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-auditor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645622 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-auditor" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645629 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="proxy-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645635 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="proxy-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645641 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefb91a4-432a-4c5d-8d28-dee413c660c3" containerName="kube-state-metrics" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645647 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefb91a4-432a-4c5d-8d28-dee413c660c3" containerName="kube-state-metrics" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645659 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerName="barbican-worker" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645666 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerName="barbican-worker" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645677 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624edf72-de15-4026-812b-36d993917176" containerName="openstack-network-exporter" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645685 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="624edf72-de15-4026-812b-36d993917176" containerName="openstack-network-exporter" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645698 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-expirer" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645705 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-expirer" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645712 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645718 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645726 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerName="barbican-keystone-listener-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645732 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerName="barbican-keystone-listener-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645741 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server-init" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645761 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server-init" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645776 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerName="extract-content" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645784 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerName="extract-content" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645794 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerName="rabbitmq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645801 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerName="rabbitmq" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645811 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-reaper" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645818 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-reaper" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645824 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9f5677-6d82-4ee0-bc2e-74c1671be521" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645830 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9f5677-6d82-4ee0-bc2e-74c1671be521" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645838 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerName="barbican-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645846 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerName="barbican-api" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645855 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb122aae-0e09-46b2-926c-037d25e79477" containerName="rabbitmq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645862 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb122aae-0e09-46b2-926c-037d25e79477" containerName="rabbitmq" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645873 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645880 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-server" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645889 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-replicator" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645896 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-replicator" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645907 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerName="barbican-keystone-listener" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645915 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerName="barbican-keystone-listener" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645922 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="sg-core" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645930 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="sg-core" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645944 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-auditor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645951 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-auditor" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645964 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="swift-recon-cron" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645971 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="swift-recon-cron" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.645982 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac086fdf-9a1a-40ab-ab2f-8a33dae28290" containerName="nova-scheduler-scheduler" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.645990 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac086fdf-9a1a-40ab-ab2f-8a33dae28290" containerName="nova-scheduler-scheduler" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646002 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerName="placement-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646009 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerName="placement-api" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646018 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646026 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646033 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerName="probe" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646040 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerName="probe" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646048 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerName="barbican-keystone-listener" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646054 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerName="barbican-keystone-listener" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646065 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646072 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-server" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646084 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51493fc6-cc57-4bbf-a6a8-0864160ade3f" containerName="nova-cell1-conductor-conductor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646091 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="51493fc6-cc57-4bbf-a6a8-0864160ade3f" containerName="nova-cell1-conductor-conductor" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646103 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452c347f-4ee3-46de-ba8f-c83300966f5d" containerName="galera" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646111 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="452c347f-4ee3-46de-ba8f-c83300966f5d" containerName="galera" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646120 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerName="barbican-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646127 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerName="barbican-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646136 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" containerName="neutron-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646143 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" containerName="neutron-api" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646157 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646164 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-api" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646176 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="rsync" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646183 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="rsync" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646196 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646203 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646213 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" containerName="memcached" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646220 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" containerName="memcached" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646229 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerName="barbican-worker" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646237 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerName="barbican-worker" Nov 28 15:51:54 crc kubenswrapper[4805]: E1128 15:51:54.646245 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerName="glance-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646251 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerName="glance-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646480 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-auditor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646493 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646505 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd6fbaf-5702-415e-8a1d-85d2c4792031" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646517 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646528 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerName="glance-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646542 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="sg-core" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646557 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac086fdf-9a1a-40ab-ab2f-8a33dae28290" containerName="nova-scheduler-scheduler" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646565 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerName="placement-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646580 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-replicator" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646587 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1db4967f-554e-4336-913a-fcebe420cf0e" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646596 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovs-vswitchd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646608 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerName="barbican-worker-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646618 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f912081e-175b-410c-bbfa-daa3a71e2179" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646626 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646638 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d7a9256-533a-4985-9a62-9bbe76af9e54" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646645 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerName="barbican-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646656 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerName="barbican-worker-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646664 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="796e4fe6-8ab3-4945-9b2b-ab4a2c99e8c5" containerName="registry-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646678 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-expirer" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646690 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646700 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646708 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="659e741d-dc3f-4463-b368-075d37a3ef0f" containerName="barbican-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646721 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-reaper" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646734 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e151fc-3a1e-4b10-8a6a-59bbb17b8d29" containerName="memcached" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646750 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e72cc62-edd5-4d0c-89a8-099d6b07e1e6" containerName="placement-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646760 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb122aae-0e09-46b2-926c-037d25e79477" containerName="rabbitmq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646771 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="abfd499c-e83b-4616-a80c-29a7e8e750e2" containerName="barbican-worker" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646783 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" containerName="neutron-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646794 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="624edf72-de15-4026-812b-36d993917176" containerName="openstack-network-exporter" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646804 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b51e8d9-5cd6-42f7-9fbf-ff2c6d856c7a" containerName="rabbitmq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646815 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7348256-32de-4696-9637-96d7323831ed" containerName="nova-metadata-metadata" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646825 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-updater" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646831 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="624edf72-de15-4026-812b-36d993917176" containerName="ovn-northd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646845 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="object-replicator" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646856 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf2e5369-a11b-4150-a291-4aeab9724d82" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646866 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8724692d-29b7-46c8-b0dd-802e570050a7" containerName="neutron-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646875 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerName="barbican-keystone-listener" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646888 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646899 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="rsync" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646911 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="95167e93-42b8-4f5f-b5a9-587b9b854f12" containerName="barbican-worker" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646923 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="89d793e5-c2d1-4630-95ff-615cd30b5d04" containerName="ovsdb-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646930 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-replicator" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646940 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="452c347f-4ee3-46de-ba8f-c83300966f5d" containerName="galera" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646949 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="ceilometer-central-agent" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646959 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="cffd57cb-2509-4a9e-8e5d-1750e4b0493e" containerName="ovn-controller" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646970 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerName="glance-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646983 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c4171ee-19b0-44c4-8e98-d14e11b9e1a6" containerName="barbican-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.646991 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2849f80f-ed31-4c7f-8f65-1132aa35b6e3" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647003 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="account-auditor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647015 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerName="cinder-scheduler" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647024 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerName="barbican-keystone-listener" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647036 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-auditor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647049 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="be5d0310-fb90-4d73-94f3-57b105a2a408" containerName="cinder-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647062 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-server" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647075 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="aefb91a4-432a-4c5d-8d28-dee413c660c3" containerName="kube-state-metrics" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647084 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c021413-239d-4ba0-979d-1c4b25d5093c" containerName="glance-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647096 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5243b85f-7ec0-4cca-9f10-de8b40b0a0f4" containerName="galera" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647107 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="container-updater" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647119 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ef711de-53e6-4705-8d2f-469b6dc2d4de" containerName="nova-api-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647130 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d69b1a-cf80-424f-ac0c-7815e7ba0375" containerName="swift-recon-cron" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647142 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbc55e90-2c22-4329-b3cc-375a68e6fe26" containerName="barbican-keystone-listener-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647156 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ef9cc1-28ab-4014-897c-9679f3d36443" containerName="glance-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647166 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="ceilometer-notification-agent" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647175 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ff5706-18a9-4e71-8289-e3d9e5d9af23" containerName="probe" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647185 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="51493fc6-cc57-4bbf-a6a8-0864160ade3f" containerName="nova-cell1-conductor-conductor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647199 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="614a68ec-3129-413f-abb1-40a73ad9137e" containerName="keystone-api" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647212 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd26b960-0994-4bb3-aef3-035519e32420" containerName="proxy-httpd" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647222 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9f5677-6d82-4ee0-bc2e-74c1671be521" containerName="mariadb-account-delete" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647233 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a363e2d1-0a91-4578-b9e4-b736b7931b03" containerName="barbican-keystone-listener-log" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.647245 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e28781d-dbe0-4ca6-a4a3-6c160ec3ecd1" containerName="nova-cell0-conductor-conductor" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.648551 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.655454 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bstbq"] Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.845892 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-utilities\") pod \"certified-operators-bstbq\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.845948 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-catalog-content\") pod \"certified-operators-bstbq\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.846025 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjw78\" (UniqueName: \"kubernetes.io/projected/75161aa3-56d7-4dfa-8613-b60099ec291f-kube-api-access-fjw78\") pod \"certified-operators-bstbq\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.946855 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-utilities\") pod \"certified-operators-bstbq\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.946907 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-catalog-content\") pod \"certified-operators-bstbq\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.946951 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjw78\" (UniqueName: \"kubernetes.io/projected/75161aa3-56d7-4dfa-8613-b60099ec291f-kube-api-access-fjw78\") pod \"certified-operators-bstbq\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.947470 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-utilities\") pod \"certified-operators-bstbq\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.947496 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-catalog-content\") pod \"certified-operators-bstbq\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.971169 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjw78\" (UniqueName: \"kubernetes.io/projected/75161aa3-56d7-4dfa-8613-b60099ec291f-kube-api-access-fjw78\") pod \"certified-operators-bstbq\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:54 crc kubenswrapper[4805]: I1128 15:51:54.972694 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:51:55 crc kubenswrapper[4805]: I1128 15:51:55.499237 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bstbq"] Nov 28 15:51:55 crc kubenswrapper[4805]: I1128 15:51:55.903435 4805 generic.go:334] "Generic (PLEG): container finished" podID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerID="43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4" exitCode=0 Nov 28 15:51:55 crc kubenswrapper[4805]: I1128 15:51:55.903478 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bstbq" event={"ID":"75161aa3-56d7-4dfa-8613-b60099ec291f","Type":"ContainerDied","Data":"43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4"} Nov 28 15:51:55 crc kubenswrapper[4805]: I1128 15:51:55.903503 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bstbq" event={"ID":"75161aa3-56d7-4dfa-8613-b60099ec291f","Type":"ContainerStarted","Data":"8df5ac07139b6493f343b5f479df1fc87843ec8e216d9faa92aeb35c39fde62d"} Nov 28 15:51:56 crc kubenswrapper[4805]: I1128 15:51:56.913492 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bstbq" event={"ID":"75161aa3-56d7-4dfa-8613-b60099ec291f","Type":"ContainerStarted","Data":"4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb"} Nov 28 15:51:57 crc kubenswrapper[4805]: I1128 15:51:57.924447 4805 generic.go:334] "Generic (PLEG): container finished" podID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerID="4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb" exitCode=0 Nov 28 15:51:57 crc kubenswrapper[4805]: I1128 15:51:57.924495 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bstbq" event={"ID":"75161aa3-56d7-4dfa-8613-b60099ec291f","Type":"ContainerDied","Data":"4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb"} Nov 28 15:51:58 crc kubenswrapper[4805]: I1128 15:51:58.934330 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bstbq" event={"ID":"75161aa3-56d7-4dfa-8613-b60099ec291f","Type":"ContainerStarted","Data":"4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd"} Nov 28 15:51:58 crc kubenswrapper[4805]: I1128 15:51:58.955846 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bstbq" podStartSLOduration=2.317975042 podStartE2EDuration="4.9558257s" podCreationTimestamp="2025-11-28 15:51:54 +0000 UTC" firstStartedPulling="2025-11-28 15:51:55.905602727 +0000 UTC m=+1542.955394038" lastFinishedPulling="2025-11-28 15:51:58.543453345 +0000 UTC m=+1545.593244696" observedRunningTime="2025-11-28 15:51:58.954829003 +0000 UTC m=+1546.004620314" watchObservedRunningTime="2025-11-28 15:51:58.9558257 +0000 UTC m=+1546.005617011" Nov 28 15:52:01 crc kubenswrapper[4805]: I1128 15:52:01.204064 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:52:01 crc kubenswrapper[4805]: E1128 15:52:01.204767 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.035148 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rxlfr"] Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.038430 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.053025 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxlfr"] Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.061163 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-catalog-content\") pod \"redhat-marketplace-rxlfr\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.061276 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-utilities\") pod \"redhat-marketplace-rxlfr\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.061413 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdr9n\" (UniqueName: \"kubernetes.io/projected/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-kube-api-access-gdr9n\") pod \"redhat-marketplace-rxlfr\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.162748 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-utilities\") pod \"redhat-marketplace-rxlfr\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.162841 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdr9n\" (UniqueName: \"kubernetes.io/projected/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-kube-api-access-gdr9n\") pod \"redhat-marketplace-rxlfr\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.163426 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-catalog-content\") pod \"redhat-marketplace-rxlfr\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.164183 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-utilities\") pod \"redhat-marketplace-rxlfr\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.164230 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-catalog-content\") pod \"redhat-marketplace-rxlfr\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.197861 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdr9n\" (UniqueName: \"kubernetes.io/projected/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-kube-api-access-gdr9n\") pod \"redhat-marketplace-rxlfr\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.376025 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.849500 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxlfr"] Nov 28 15:52:02 crc kubenswrapper[4805]: I1128 15:52:02.966888 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxlfr" event={"ID":"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7","Type":"ContainerStarted","Data":"fb68754de86dd2dde1ded00b000e6344726321a615df6793afc5fa3cdfb130bd"} Nov 28 15:52:03 crc kubenswrapper[4805]: I1128 15:52:03.978404 4805 generic.go:334] "Generic (PLEG): container finished" podID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerID="dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63" exitCode=0 Nov 28 15:52:03 crc kubenswrapper[4805]: I1128 15:52:03.978477 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxlfr" event={"ID":"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7","Type":"ContainerDied","Data":"dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63"} Nov 28 15:52:04 crc kubenswrapper[4805]: I1128 15:52:04.973578 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:52:04 crc kubenswrapper[4805]: I1128 15:52:04.973630 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:52:04 crc kubenswrapper[4805]: I1128 15:52:04.987893 4805 generic.go:334] "Generic (PLEG): container finished" podID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerID="25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f" exitCode=0 Nov 28 15:52:04 crc kubenswrapper[4805]: I1128 15:52:04.987957 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxlfr" event={"ID":"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7","Type":"ContainerDied","Data":"25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f"} Nov 28 15:52:05 crc kubenswrapper[4805]: I1128 15:52:05.037588 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:52:05 crc kubenswrapper[4805]: I1128 15:52:05.091860 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:52:06 crc kubenswrapper[4805]: I1128 15:52:06.002219 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxlfr" event={"ID":"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7","Type":"ContainerStarted","Data":"7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428"} Nov 28 15:52:06 crc kubenswrapper[4805]: I1128 15:52:06.022125 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rxlfr" podStartSLOduration=2.609291557 podStartE2EDuration="4.022098334s" podCreationTimestamp="2025-11-28 15:52:02 +0000 UTC" firstStartedPulling="2025-11-28 15:52:03.980108335 +0000 UTC m=+1551.029899686" lastFinishedPulling="2025-11-28 15:52:05.392915152 +0000 UTC m=+1552.442706463" observedRunningTime="2025-11-28 15:52:06.018856136 +0000 UTC m=+1553.068647437" watchObservedRunningTime="2025-11-28 15:52:06.022098334 +0000 UTC m=+1553.071889665" Nov 28 15:52:07 crc kubenswrapper[4805]: I1128 15:52:07.418624 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bstbq"] Nov 28 15:52:07 crc kubenswrapper[4805]: I1128 15:52:07.419078 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bstbq" podUID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerName="registry-server" containerID="cri-o://4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd" gracePeriod=2 Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.311074 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.456502 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjw78\" (UniqueName: \"kubernetes.io/projected/75161aa3-56d7-4dfa-8613-b60099ec291f-kube-api-access-fjw78\") pod \"75161aa3-56d7-4dfa-8613-b60099ec291f\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.456568 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-utilities\") pod \"75161aa3-56d7-4dfa-8613-b60099ec291f\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.456697 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-catalog-content\") pod \"75161aa3-56d7-4dfa-8613-b60099ec291f\" (UID: \"75161aa3-56d7-4dfa-8613-b60099ec291f\") " Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.458270 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-utilities" (OuterVolumeSpecName: "utilities") pod "75161aa3-56d7-4dfa-8613-b60099ec291f" (UID: "75161aa3-56d7-4dfa-8613-b60099ec291f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.464435 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75161aa3-56d7-4dfa-8613-b60099ec291f-kube-api-access-fjw78" (OuterVolumeSpecName: "kube-api-access-fjw78") pod "75161aa3-56d7-4dfa-8613-b60099ec291f" (UID: "75161aa3-56d7-4dfa-8613-b60099ec291f"). InnerVolumeSpecName "kube-api-access-fjw78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.516558 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75161aa3-56d7-4dfa-8613-b60099ec291f" (UID: "75161aa3-56d7-4dfa-8613-b60099ec291f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.558333 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjw78\" (UniqueName: \"kubernetes.io/projected/75161aa3-56d7-4dfa-8613-b60099ec291f-kube-api-access-fjw78\") on node \"crc\" DevicePath \"\"" Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.558408 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:52:08 crc kubenswrapper[4805]: I1128 15:52:08.558429 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75161aa3-56d7-4dfa-8613-b60099ec291f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.026210 4805 generic.go:334] "Generic (PLEG): container finished" podID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerID="4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd" exitCode=0 Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.026258 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bstbq" event={"ID":"75161aa3-56d7-4dfa-8613-b60099ec291f","Type":"ContainerDied","Data":"4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd"} Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.026284 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bstbq" event={"ID":"75161aa3-56d7-4dfa-8613-b60099ec291f","Type":"ContainerDied","Data":"8df5ac07139b6493f343b5f479df1fc87843ec8e216d9faa92aeb35c39fde62d"} Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.026301 4805 scope.go:117] "RemoveContainer" containerID="4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.026425 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bstbq" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.046119 4805 scope.go:117] "RemoveContainer" containerID="4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.056341 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bstbq"] Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.062371 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bstbq"] Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.083713 4805 scope.go:117] "RemoveContainer" containerID="43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.099300 4805 scope.go:117] "RemoveContainer" containerID="4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd" Nov 28 15:52:09 crc kubenswrapper[4805]: E1128 15:52:09.099720 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd\": container with ID starting with 4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd not found: ID does not exist" containerID="4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.099752 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd"} err="failed to get container status \"4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd\": rpc error: code = NotFound desc = could not find container \"4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd\": container with ID starting with 4d0447d0a2619a807aeaac3c399e6f16b28ca5bfe391688bfa73bc78e8b787fd not found: ID does not exist" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.099773 4805 scope.go:117] "RemoveContainer" containerID="4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb" Nov 28 15:52:09 crc kubenswrapper[4805]: E1128 15:52:09.100043 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb\": container with ID starting with 4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb not found: ID does not exist" containerID="4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.100066 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb"} err="failed to get container status \"4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb\": rpc error: code = NotFound desc = could not find container \"4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb\": container with ID starting with 4c9ce3d24859bc03522585587975920a573e59f8037525baf3e9855f01a3b6eb not found: ID does not exist" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.100082 4805 scope.go:117] "RemoveContainer" containerID="43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4" Nov 28 15:52:09 crc kubenswrapper[4805]: E1128 15:52:09.100321 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4\": container with ID starting with 43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4 not found: ID does not exist" containerID="43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.100338 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4"} err="failed to get container status \"43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4\": rpc error: code = NotFound desc = could not find container \"43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4\": container with ID starting with 43c16ebcfecdb4b988477920bc3022d6859ae04adda6b793450e8fd77249e3d4 not found: ID does not exist" Nov 28 15:52:09 crc kubenswrapper[4805]: I1128 15:52:09.213689 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75161aa3-56d7-4dfa-8613-b60099ec291f" path="/var/lib/kubelet/pods/75161aa3-56d7-4dfa-8613-b60099ec291f/volumes" Nov 28 15:52:12 crc kubenswrapper[4805]: I1128 15:52:12.376629 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:12 crc kubenswrapper[4805]: I1128 15:52:12.377265 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:12 crc kubenswrapper[4805]: I1128 15:52:12.437164 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:13 crc kubenswrapper[4805]: I1128 15:52:13.146335 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:13 crc kubenswrapper[4805]: I1128 15:52:13.285880 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxlfr"] Nov 28 15:52:14 crc kubenswrapper[4805]: I1128 15:52:14.204075 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:52:14 crc kubenswrapper[4805]: E1128 15:52:14.204301 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:52:15 crc kubenswrapper[4805]: I1128 15:52:15.082626 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rxlfr" podUID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerName="registry-server" containerID="cri-o://7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428" gracePeriod=2 Nov 28 15:52:15 crc kubenswrapper[4805]: I1128 15:52:15.988792 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.091745 4805 generic.go:334] "Generic (PLEG): container finished" podID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerID="7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428" exitCode=0 Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.091797 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxlfr" event={"ID":"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7","Type":"ContainerDied","Data":"7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428"} Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.091829 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxlfr" event={"ID":"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7","Type":"ContainerDied","Data":"fb68754de86dd2dde1ded00b000e6344726321a615df6793afc5fa3cdfb130bd"} Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.091849 4805 scope.go:117] "RemoveContainer" containerID="7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.091992 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxlfr" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.108534 4805 scope.go:117] "RemoveContainer" containerID="25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.128681 4805 scope.go:117] "RemoveContainer" containerID="dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.158532 4805 scope.go:117] "RemoveContainer" containerID="7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428" Nov 28 15:52:16 crc kubenswrapper[4805]: E1128 15:52:16.159122 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428\": container with ID starting with 7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428 not found: ID does not exist" containerID="7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.159202 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428"} err="failed to get container status \"7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428\": rpc error: code = NotFound desc = could not find container \"7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428\": container with ID starting with 7fe4c6e9f1c173b99cd01aa9cb11a9e4196f8c465d64c182d223843319e9e428 not found: ID does not exist" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.159242 4805 scope.go:117] "RemoveContainer" containerID="25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f" Nov 28 15:52:16 crc kubenswrapper[4805]: E1128 15:52:16.159838 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f\": container with ID starting with 25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f not found: ID does not exist" containerID="25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.159872 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f"} err="failed to get container status \"25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f\": rpc error: code = NotFound desc = could not find container \"25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f\": container with ID starting with 25a4a4a5285e2df8d674464a071666ac6f1e83512545a4988495a95de3535a9f not found: ID does not exist" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.159898 4805 scope.go:117] "RemoveContainer" containerID="dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63" Nov 28 15:52:16 crc kubenswrapper[4805]: E1128 15:52:16.160295 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63\": container with ID starting with dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63 not found: ID does not exist" containerID="dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.160355 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63"} err="failed to get container status \"dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63\": rpc error: code = NotFound desc = could not find container \"dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63\": container with ID starting with dce683be45f611bd73f0f13a7361c71b251a4fab6e3807bd70d8d40c911ecb63 not found: ID does not exist" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.179110 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-utilities\") pod \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.179215 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-catalog-content\") pod \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.179253 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdr9n\" (UniqueName: \"kubernetes.io/projected/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-kube-api-access-gdr9n\") pod \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\" (UID: \"a6da7c62-0cec-420f-a64b-e5ebe6fb38b7\") " Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.180487 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-utilities" (OuterVolumeSpecName: "utilities") pod "a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" (UID: "a6da7c62-0cec-420f-a64b-e5ebe6fb38b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.185286 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-kube-api-access-gdr9n" (OuterVolumeSpecName: "kube-api-access-gdr9n") pod "a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" (UID: "a6da7c62-0cec-420f-a64b-e5ebe6fb38b7"). InnerVolumeSpecName "kube-api-access-gdr9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.198174 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" (UID: "a6da7c62-0cec-420f-a64b-e5ebe6fb38b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.281412 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.281450 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.281460 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdr9n\" (UniqueName: \"kubernetes.io/projected/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7-kube-api-access-gdr9n\") on node \"crc\" DevicePath \"\"" Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.423543 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxlfr"] Nov 28 15:52:16 crc kubenswrapper[4805]: I1128 15:52:16.433004 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxlfr"] Nov 28 15:52:17 crc kubenswrapper[4805]: I1128 15:52:17.223525 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" path="/var/lib/kubelet/pods/a6da7c62-0cec-420f-a64b-e5ebe6fb38b7/volumes" Nov 28 15:52:26 crc kubenswrapper[4805]: I1128 15:52:26.204013 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:52:26 crc kubenswrapper[4805]: E1128 15:52:26.204421 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.835990 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r528f"] Nov 28 15:52:33 crc kubenswrapper[4805]: E1128 15:52:33.837002 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerName="extract-utilities" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.837022 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerName="extract-utilities" Nov 28 15:52:33 crc kubenswrapper[4805]: E1128 15:52:33.837048 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerName="extract-utilities" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.837059 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerName="extract-utilities" Nov 28 15:52:33 crc kubenswrapper[4805]: E1128 15:52:33.837088 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerName="registry-server" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.837100 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerName="registry-server" Nov 28 15:52:33 crc kubenswrapper[4805]: E1128 15:52:33.837120 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerName="registry-server" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.837131 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerName="registry-server" Nov 28 15:52:33 crc kubenswrapper[4805]: E1128 15:52:33.837165 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerName="extract-content" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.837177 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerName="extract-content" Nov 28 15:52:33 crc kubenswrapper[4805]: E1128 15:52:33.837203 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerName="extract-content" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.837214 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerName="extract-content" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.837459 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="75161aa3-56d7-4dfa-8613-b60099ec291f" containerName="registry-server" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.837512 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6da7c62-0cec-420f-a64b-e5ebe6fb38b7" containerName="registry-server" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.839191 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:33 crc kubenswrapper[4805]: I1128 15:52:33.862944 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r528f"] Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.024429 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-catalog-content\") pod \"community-operators-r528f\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.024568 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-utilities\") pod \"community-operators-r528f\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.024740 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tk8h\" (UniqueName: \"kubernetes.io/projected/bfc94d69-d8b7-4edc-816f-370315f47a1e-kube-api-access-4tk8h\") pod \"community-operators-r528f\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.126025 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-catalog-content\") pod \"community-operators-r528f\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.126097 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-utilities\") pod \"community-operators-r528f\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.126157 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tk8h\" (UniqueName: \"kubernetes.io/projected/bfc94d69-d8b7-4edc-816f-370315f47a1e-kube-api-access-4tk8h\") pod \"community-operators-r528f\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.126582 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-utilities\") pod \"community-operators-r528f\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.126618 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-catalog-content\") pod \"community-operators-r528f\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.148667 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tk8h\" (UniqueName: \"kubernetes.io/projected/bfc94d69-d8b7-4edc-816f-370315f47a1e-kube-api-access-4tk8h\") pod \"community-operators-r528f\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.185475 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:34 crc kubenswrapper[4805]: I1128 15:52:34.677019 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r528f"] Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.264016 4805 generic.go:334] "Generic (PLEG): container finished" podID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerID="6e83d179f50eac5a2a7387fbc77da1f2a45e4d83fb3a7c45fac7c92e630d5c34" exitCode=0 Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.264102 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r528f" event={"ID":"bfc94d69-d8b7-4edc-816f-370315f47a1e","Type":"ContainerDied","Data":"6e83d179f50eac5a2a7387fbc77da1f2a45e4d83fb3a7c45fac7c92e630d5c34"} Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.266801 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r528f" event={"ID":"bfc94d69-d8b7-4edc-816f-370315f47a1e","Type":"ContainerStarted","Data":"342b1f62dbf0c1936fe1eb8ba7775b61a4f375288e9b852d89a11e06fee60c1f"} Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.531854 4805 scope.go:117] "RemoveContainer" containerID="fe684bfa9591b55be6b83497c5f2059493fe4abdfa52867aaec588e8310d8893" Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.553209 4805 scope.go:117] "RemoveContainer" containerID="02772248390f9aa46863e063f9eac9961344ec1704c67c8cfde40db99e473634" Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.571305 4805 scope.go:117] "RemoveContainer" containerID="66fc00b5a0c7d5310a8c3fbb5a766ab2e30e4c1049270d558c221c8c94d6292e" Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.586594 4805 scope.go:117] "RemoveContainer" containerID="8bdacf735cc55e4eb9afb66b2e37c11f6d54e6bb6bec04837719d6ce7e4e5af8" Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.631315 4805 scope.go:117] "RemoveContainer" containerID="6260afcad67d63a4d5d64b9acce81ba750c382607d4f5264abef41c12d820237" Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.652944 4805 scope.go:117] "RemoveContainer" containerID="3662b2162435ede3ed240fafeef413deeef6d00aecf000c0f5887352f3c5a777" Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.676259 4805 scope.go:117] "RemoveContainer" containerID="65df2285b683dab8a99695a41df58b3ed9b436df295bd383c9af5720935f1dbb" Nov 28 15:52:35 crc kubenswrapper[4805]: I1128 15:52:35.706324 4805 scope.go:117] "RemoveContainer" containerID="14da7f9c8fb9e23ab79dd5e17973e13aa4fe285d149b8b95925f3a3f92426818" Nov 28 15:52:36 crc kubenswrapper[4805]: I1128 15:52:36.285431 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r528f" event={"ID":"bfc94d69-d8b7-4edc-816f-370315f47a1e","Type":"ContainerStarted","Data":"8b5603ff7aaf16d9cf7ac29a8997cb074627dd2634d2ab1608705f2669189f4c"} Nov 28 15:52:37 crc kubenswrapper[4805]: I1128 15:52:37.293831 4805 generic.go:334] "Generic (PLEG): container finished" podID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerID="8b5603ff7aaf16d9cf7ac29a8997cb074627dd2634d2ab1608705f2669189f4c" exitCode=0 Nov 28 15:52:37 crc kubenswrapper[4805]: I1128 15:52:37.293881 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r528f" event={"ID":"bfc94d69-d8b7-4edc-816f-370315f47a1e","Type":"ContainerDied","Data":"8b5603ff7aaf16d9cf7ac29a8997cb074627dd2634d2ab1608705f2669189f4c"} Nov 28 15:52:37 crc kubenswrapper[4805]: I1128 15:52:37.293911 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r528f" event={"ID":"bfc94d69-d8b7-4edc-816f-370315f47a1e","Type":"ContainerStarted","Data":"38d7aca40c547ca0470638a9f0e9dcc861810d6f11c574d70467d15fee0891a8"} Nov 28 15:52:37 crc kubenswrapper[4805]: I1128 15:52:37.312454 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r528f" podStartSLOduration=2.487974824 podStartE2EDuration="4.312435494s" podCreationTimestamp="2025-11-28 15:52:33 +0000 UTC" firstStartedPulling="2025-11-28 15:52:35.26730569 +0000 UTC m=+1582.317097001" lastFinishedPulling="2025-11-28 15:52:37.09176636 +0000 UTC m=+1584.141557671" observedRunningTime="2025-11-28 15:52:37.308371192 +0000 UTC m=+1584.358162523" watchObservedRunningTime="2025-11-28 15:52:37.312435494 +0000 UTC m=+1584.362226805" Nov 28 15:52:38 crc kubenswrapper[4805]: I1128 15:52:38.203791 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:52:38 crc kubenswrapper[4805]: E1128 15:52:38.204128 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:52:44 crc kubenswrapper[4805]: I1128 15:52:44.186263 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:44 crc kubenswrapper[4805]: I1128 15:52:44.186786 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:44 crc kubenswrapper[4805]: I1128 15:52:44.257299 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:44 crc kubenswrapper[4805]: I1128 15:52:44.408432 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:44 crc kubenswrapper[4805]: I1128 15:52:44.490251 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r528f"] Nov 28 15:52:46 crc kubenswrapper[4805]: I1128 15:52:46.372844 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r528f" podUID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerName="registry-server" containerID="cri-o://38d7aca40c547ca0470638a9f0e9dcc861810d6f11c574d70467d15fee0891a8" gracePeriod=2 Nov 28 15:52:47 crc kubenswrapper[4805]: I1128 15:52:47.384693 4805 generic.go:334] "Generic (PLEG): container finished" podID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerID="38d7aca40c547ca0470638a9f0e9dcc861810d6f11c574d70467d15fee0891a8" exitCode=0 Nov 28 15:52:47 crc kubenswrapper[4805]: I1128 15:52:47.384758 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r528f" event={"ID":"bfc94d69-d8b7-4edc-816f-370315f47a1e","Type":"ContainerDied","Data":"38d7aca40c547ca0470638a9f0e9dcc861810d6f11c574d70467d15fee0891a8"} Nov 28 15:52:47 crc kubenswrapper[4805]: I1128 15:52:47.943564 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.142140 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-catalog-content\") pod \"bfc94d69-d8b7-4edc-816f-370315f47a1e\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.142290 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tk8h\" (UniqueName: \"kubernetes.io/projected/bfc94d69-d8b7-4edc-816f-370315f47a1e-kube-api-access-4tk8h\") pod \"bfc94d69-d8b7-4edc-816f-370315f47a1e\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.142471 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-utilities\") pod \"bfc94d69-d8b7-4edc-816f-370315f47a1e\" (UID: \"bfc94d69-d8b7-4edc-816f-370315f47a1e\") " Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.146071 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-utilities" (OuterVolumeSpecName: "utilities") pod "bfc94d69-d8b7-4edc-816f-370315f47a1e" (UID: "bfc94d69-d8b7-4edc-816f-370315f47a1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.156702 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfc94d69-d8b7-4edc-816f-370315f47a1e-kube-api-access-4tk8h" (OuterVolumeSpecName: "kube-api-access-4tk8h") pod "bfc94d69-d8b7-4edc-816f-370315f47a1e" (UID: "bfc94d69-d8b7-4edc-816f-370315f47a1e"). InnerVolumeSpecName "kube-api-access-4tk8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.208395 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfc94d69-d8b7-4edc-816f-370315f47a1e" (UID: "bfc94d69-d8b7-4edc-816f-370315f47a1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.245974 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tk8h\" (UniqueName: \"kubernetes.io/projected/bfc94d69-d8b7-4edc-816f-370315f47a1e-kube-api-access-4tk8h\") on node \"crc\" DevicePath \"\"" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.246035 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.246052 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc94d69-d8b7-4edc-816f-370315f47a1e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.395109 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r528f" event={"ID":"bfc94d69-d8b7-4edc-816f-370315f47a1e","Type":"ContainerDied","Data":"342b1f62dbf0c1936fe1eb8ba7775b61a4f375288e9b852d89a11e06fee60c1f"} Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.396329 4805 scope.go:117] "RemoveContainer" containerID="38d7aca40c547ca0470638a9f0e9dcc861810d6f11c574d70467d15fee0891a8" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.395267 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r528f" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.423603 4805 scope.go:117] "RemoveContainer" containerID="8b5603ff7aaf16d9cf7ac29a8997cb074627dd2634d2ab1608705f2669189f4c" Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.439413 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r528f"] Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.446618 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r528f"] Nov 28 15:52:48 crc kubenswrapper[4805]: I1128 15:52:48.457377 4805 scope.go:117] "RemoveContainer" containerID="6e83d179f50eac5a2a7387fbc77da1f2a45e4d83fb3a7c45fac7c92e630d5c34" Nov 28 15:52:49 crc kubenswrapper[4805]: I1128 15:52:49.214225 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfc94d69-d8b7-4edc-816f-370315f47a1e" path="/var/lib/kubelet/pods/bfc94d69-d8b7-4edc-816f-370315f47a1e/volumes" Nov 28 15:52:53 crc kubenswrapper[4805]: I1128 15:52:53.213894 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:52:53 crc kubenswrapper[4805]: E1128 15:52:53.215781 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:53:06 crc kubenswrapper[4805]: I1128 15:53:06.203912 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:53:06 crc kubenswrapper[4805]: E1128 15:53:06.204664 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:53:19 crc kubenswrapper[4805]: I1128 15:53:19.203858 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:53:19 crc kubenswrapper[4805]: E1128 15:53:19.204742 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:53:33 crc kubenswrapper[4805]: I1128 15:53:33.212672 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:53:33 crc kubenswrapper[4805]: E1128 15:53:33.214027 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:53:35 crc kubenswrapper[4805]: I1128 15:53:35.868905 4805 scope.go:117] "RemoveContainer" containerID="ab13ce4cb62ee031c9d0c066d028e77475c256f8b5848622a9fddf54b469b612" Nov 28 15:53:35 crc kubenswrapper[4805]: I1128 15:53:35.890017 4805 scope.go:117] "RemoveContainer" containerID="b6992038b195926e6c2ae3d4f1f5dd8116283fe079e24263ee47fb35fb22332a" Nov 28 15:53:35 crc kubenswrapper[4805]: I1128 15:53:35.955861 4805 scope.go:117] "RemoveContainer" containerID="fe1f555311c706978cf71b633f58b4d74d761a478727bbea477a1a6d7694312a" Nov 28 15:53:35 crc kubenswrapper[4805]: I1128 15:53:35.988800 4805 scope.go:117] "RemoveContainer" containerID="23ec9bca9f3899799fc7a6bd1a69fcbc436496ef980bbce600934b76418db1a3" Nov 28 15:53:36 crc kubenswrapper[4805]: I1128 15:53:36.013391 4805 scope.go:117] "RemoveContainer" containerID="52836939c88b31b81ae1b8b1369c201d032ad656c44f7c7999db8fd5105ccd6a" Nov 28 15:53:36 crc kubenswrapper[4805]: I1128 15:53:36.039510 4805 scope.go:117] "RemoveContainer" containerID="eb8fb64ef7fabdfff26fc63df7f5afd8cdf3b2709521bf51427cfd02832aac89" Nov 28 15:53:36 crc kubenswrapper[4805]: I1128 15:53:36.063887 4805 scope.go:117] "RemoveContainer" containerID="f8566ba278f7725c827f682fa40e5bc7af932461753c88f412a5530ab9a78ea3" Nov 28 15:53:46 crc kubenswrapper[4805]: I1128 15:53:46.204820 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:53:46 crc kubenswrapper[4805]: E1128 15:53:46.205572 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:53:59 crc kubenswrapper[4805]: I1128 15:53:59.205126 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:53:59 crc kubenswrapper[4805]: E1128 15:53:59.206435 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:54:14 crc kubenswrapper[4805]: I1128 15:54:14.204861 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:54:14 crc kubenswrapper[4805]: E1128 15:54:14.205887 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:54:28 crc kubenswrapper[4805]: I1128 15:54:28.204876 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:54:28 crc kubenswrapper[4805]: E1128 15:54:28.207086 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:54:36 crc kubenswrapper[4805]: I1128 15:54:36.173522 4805 scope.go:117] "RemoveContainer" containerID="af248f060f9652198f3fb50557ebfd252fb89de3f39e24268855915e1ebf72bd" Nov 28 15:54:36 crc kubenswrapper[4805]: I1128 15:54:36.192766 4805 scope.go:117] "RemoveContainer" containerID="93594d7686a712e66ead92d80c77d078b1f2f43f80e381c3d84546645e2f43fc" Nov 28 15:54:36 crc kubenswrapper[4805]: I1128 15:54:36.238010 4805 scope.go:117] "RemoveContainer" containerID="c8dee08d7ace9297ae12c4f0d44cad8e080785d18e73355b29ebf5cc4b33729a" Nov 28 15:54:43 crc kubenswrapper[4805]: I1128 15:54:43.208760 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:54:43 crc kubenswrapper[4805]: E1128 15:54:43.209505 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:54:54 crc kubenswrapper[4805]: I1128 15:54:54.203892 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:54:54 crc kubenswrapper[4805]: E1128 15:54:54.204677 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:55:05 crc kubenswrapper[4805]: I1128 15:55:05.204606 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:55:05 crc kubenswrapper[4805]: E1128 15:55:05.205337 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:55:16 crc kubenswrapper[4805]: I1128 15:55:16.204542 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:55:16 crc kubenswrapper[4805]: E1128 15:55:16.205496 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.648456 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-stv9f/must-gather-lgp5j"] Nov 28 15:55:22 crc kubenswrapper[4805]: E1128 15:55:22.649588 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerName="registry-server" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.649613 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerName="registry-server" Nov 28 15:55:22 crc kubenswrapper[4805]: E1128 15:55:22.649635 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerName="extract-content" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.649641 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerName="extract-content" Nov 28 15:55:22 crc kubenswrapper[4805]: E1128 15:55:22.649664 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerName="extract-utilities" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.649671 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerName="extract-utilities" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.649987 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfc94d69-d8b7-4edc-816f-370315f47a1e" containerName="registry-server" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.651941 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.679599 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-stv9f"/"openshift-service-ca.crt" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.679977 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-stv9f"/"kube-root-ca.crt" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.700410 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-stv9f/must-gather-lgp5j"] Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.843238 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9cbde7b-d74e-4257-beb9-3c1557b5029b-must-gather-output\") pod \"must-gather-lgp5j\" (UID: \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\") " pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.843334 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8w8z\" (UniqueName: \"kubernetes.io/projected/e9cbde7b-d74e-4257-beb9-3c1557b5029b-kube-api-access-m8w8z\") pod \"must-gather-lgp5j\" (UID: \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\") " pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.944550 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9cbde7b-d74e-4257-beb9-3c1557b5029b-must-gather-output\") pod \"must-gather-lgp5j\" (UID: \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\") " pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.944627 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8w8z\" (UniqueName: \"kubernetes.io/projected/e9cbde7b-d74e-4257-beb9-3c1557b5029b-kube-api-access-m8w8z\") pod \"must-gather-lgp5j\" (UID: \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\") " pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.945029 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9cbde7b-d74e-4257-beb9-3c1557b5029b-must-gather-output\") pod \"must-gather-lgp5j\" (UID: \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\") " pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:55:22 crc kubenswrapper[4805]: I1128 15:55:22.965339 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8w8z\" (UniqueName: \"kubernetes.io/projected/e9cbde7b-d74e-4257-beb9-3c1557b5029b-kube-api-access-m8w8z\") pod \"must-gather-lgp5j\" (UID: \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\") " pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:55:23 crc kubenswrapper[4805]: I1128 15:55:23.020987 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:55:23 crc kubenswrapper[4805]: I1128 15:55:23.441220 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-stv9f/must-gather-lgp5j"] Nov 28 15:55:23 crc kubenswrapper[4805]: I1128 15:55:23.442963 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:55:24 crc kubenswrapper[4805]: I1128 15:55:24.338714 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-stv9f/must-gather-lgp5j" event={"ID":"e9cbde7b-d74e-4257-beb9-3c1557b5029b","Type":"ContainerStarted","Data":"3f939d44ad04648e9733b49427d79166cef7089c5e877fdeb33369567e9ce2a9"} Nov 28 15:55:28 crc kubenswrapper[4805]: I1128 15:55:28.372977 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-stv9f/must-gather-lgp5j" event={"ID":"e9cbde7b-d74e-4257-beb9-3c1557b5029b","Type":"ContainerStarted","Data":"5a3f74e6b9a054b6fde8b003540ccf082ffdabe0f44e766655ee750442c24b12"} Nov 28 15:55:28 crc kubenswrapper[4805]: I1128 15:55:28.373709 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-stv9f/must-gather-lgp5j" event={"ID":"e9cbde7b-d74e-4257-beb9-3c1557b5029b","Type":"ContainerStarted","Data":"178bdd4a9c60adb2c226f2c50aa447c4f9f71d938bcb0888171001944499a496"} Nov 28 15:55:28 crc kubenswrapper[4805]: I1128 15:55:28.393666 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-stv9f/must-gather-lgp5j" podStartSLOduration=2.284617514 podStartE2EDuration="6.393645905s" podCreationTimestamp="2025-11-28 15:55:22 +0000 UTC" firstStartedPulling="2025-11-28 15:55:23.442916861 +0000 UTC m=+1750.492708172" lastFinishedPulling="2025-11-28 15:55:27.551945232 +0000 UTC m=+1754.601736563" observedRunningTime="2025-11-28 15:55:28.387485397 +0000 UTC m=+1755.437276708" watchObservedRunningTime="2025-11-28 15:55:28.393645905 +0000 UTC m=+1755.443437236" Nov 28 15:55:31 crc kubenswrapper[4805]: I1128 15:55:31.203986 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:55:31 crc kubenswrapper[4805]: E1128 15:55:31.204614 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:55:36 crc kubenswrapper[4805]: I1128 15:55:36.356061 4805 scope.go:117] "RemoveContainer" containerID="a92a82d6fc0b62799e0012cf79d6be83a592db16c726700d8557140daac5fbef" Nov 28 15:55:36 crc kubenswrapper[4805]: I1128 15:55:36.380390 4805 scope.go:117] "RemoveContainer" containerID="e210a202934599ab4809d38358f193e2eb1eb77d9f8eb89fa27b616eca30451d" Nov 28 15:55:36 crc kubenswrapper[4805]: I1128 15:55:36.403965 4805 scope.go:117] "RemoveContainer" containerID="fb7714c312b5ac2a0fbfa8a0f6a228e77cc36a21c9b90183adfa3477910eb34b" Nov 28 15:55:36 crc kubenswrapper[4805]: I1128 15:55:36.429465 4805 scope.go:117] "RemoveContainer" containerID="aae5a7230edb099b422ff386a01d34c04320a5e4d35f17b3aaf74b791db82ff0" Nov 28 15:55:36 crc kubenswrapper[4805]: I1128 15:55:36.471306 4805 scope.go:117] "RemoveContainer" containerID="8af92c9e96307e97e3180f6f4e515c45e8d404b658b8fecda5e22417d08f76db" Nov 28 15:55:36 crc kubenswrapper[4805]: I1128 15:55:36.486828 4805 scope.go:117] "RemoveContainer" containerID="b53b84616761467f8383fb98e17f379ec12fbb906d9bd0cffe2c80f86e68acfb" Nov 28 15:55:36 crc kubenswrapper[4805]: I1128 15:55:36.512669 4805 scope.go:117] "RemoveContainer" containerID="5bb454b44321bc150d612a41130b3c944d8a6a48515220771451697975c341b2" Nov 28 15:55:36 crc kubenswrapper[4805]: I1128 15:55:36.534155 4805 scope.go:117] "RemoveContainer" containerID="0e6f68b4ef1f21b2f8bb9ebc77214db715f155aaed05e0a61d2a59569c636bbf" Nov 28 15:55:46 crc kubenswrapper[4805]: I1128 15:55:46.204503 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:55:46 crc kubenswrapper[4805]: E1128 15:55:46.205650 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:56:00 crc kubenswrapper[4805]: I1128 15:56:00.204043 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:56:00 crc kubenswrapper[4805]: E1128 15:56:00.205061 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9pzmp_openshift-machine-config-operator(73da70d8-a7dc-4fca-9e65-9c0d0d815966)\"" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" Nov 28 15:56:13 crc kubenswrapper[4805]: I1128 15:56:13.209768 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:56:13 crc kubenswrapper[4805]: I1128 15:56:13.670283 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"482bf7c90ddabbe0361f94c2e9d217342f00d85b78e5e55882137ac7b4c59d0d"} Nov 28 15:56:21 crc kubenswrapper[4805]: I1128 15:56:21.567427 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd_7e48a537-6a16-4037-96aa-242276b036a8/util/0.log" Nov 28 15:56:21 crc kubenswrapper[4805]: I1128 15:56:21.764272 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd_7e48a537-6a16-4037-96aa-242276b036a8/pull/0.log" Nov 28 15:56:21 crc kubenswrapper[4805]: I1128 15:56:21.768807 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd_7e48a537-6a16-4037-96aa-242276b036a8/pull/0.log" Nov 28 15:56:21 crc kubenswrapper[4805]: I1128 15:56:21.784071 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd_7e48a537-6a16-4037-96aa-242276b036a8/util/0.log" Nov 28 15:56:21 crc kubenswrapper[4805]: I1128 15:56:21.917693 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd_7e48a537-6a16-4037-96aa-242276b036a8/pull/0.log" Nov 28 15:56:21 crc kubenswrapper[4805]: I1128 15:56:21.936272 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd_7e48a537-6a16-4037-96aa-242276b036a8/util/0.log" Nov 28 15:56:21 crc kubenswrapper[4805]: I1128 15:56:21.975279 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c59326877ec41274633a1fdb87982e94535bc5dc9830a0d3dbe2c54aapr8cd_7e48a537-6a16-4037-96aa-242276b036a8/extract/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.147183 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-lh9jk_446d7f87-a4e9-4a87-9091-926345489a0d/kube-rbac-proxy/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.165095 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-lh9jk_446d7f87-a4e9-4a87-9091-926345489a0d/manager/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.202990 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-m6cln_6d7afdad-6a62-4c69-b453-dcf5b917a568/kube-rbac-proxy/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.326591 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-78jxp_697a2a64-1776-4bb4-95f4-901e1ae2bbc2/kube-rbac-proxy/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.326900 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-m6cln_6d7afdad-6a62-4c69-b453-dcf5b917a568/manager/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.377894 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-78jxp_697a2a64-1776-4bb4-95f4-901e1ae2bbc2/manager/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.512024 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-p9lc4_8ff9654c-7de4-4cc2-af76-863ac6cb8694/kube-rbac-proxy/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.615820 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-p9lc4_8ff9654c-7de4-4cc2-af76-863ac6cb8694/manager/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.716317 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-jl2bq_d786fb2a-908c-42bc-8889-f85d48a824c3/manager/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.736021 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-jl2bq_d786fb2a-908c-42bc-8889-f85d48a824c3/kube-rbac-proxy/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.829052 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-jn44r_4a95d51b-3298-4575-84d3-5b1f1174edbe/kube-rbac-proxy/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.891433 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-jn44r_4a95d51b-3298-4575-84d3-5b1f1174edbe/manager/0.log" Nov 28 15:56:22 crc kubenswrapper[4805]: I1128 15:56:22.955125 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-2p9q6_dfdaba79-b86b-4a03-87e0-ed9f712ea4bb/kube-rbac-proxy/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.091635 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-2p9q6_dfdaba79-b86b-4a03-87e0-ed9f712ea4bb/manager/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.146094 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-v7f97_f0bf9b46-6d36-47b4-bbac-374d79b553fc/manager/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.148857 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-v7f97_f0bf9b46-6d36-47b4-bbac-374d79b553fc/kube-rbac-proxy/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.265922 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-5m48r_cd24b067-2061-47eb-8494-5a531f13b111/kube-rbac-proxy/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.360911 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-5m48r_cd24b067-2061-47eb-8494-5a531f13b111/manager/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.449066 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-bqmpw_3bad1d97-f824-4616-a08d-9e00fd2c5201/kube-rbac-proxy/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.482482 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-bqmpw_3bad1d97-f824-4616-a08d-9e00fd2c5201/manager/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.552518 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-skdkr_49986953-d4ca-4e5f-ac36-f91d86098ff5/kube-rbac-proxy/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.673832 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-skdkr_49986953-d4ca-4e5f-ac36-f91d86098ff5/manager/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.733916 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-lkqjk_75b25042-ef68-47da-873b-6750083ecc7e/kube-rbac-proxy/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.798533 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-lkqjk_75b25042-ef68-47da-873b-6750083ecc7e/manager/0.log" Nov 28 15:56:23 crc kubenswrapper[4805]: I1128 15:56:23.903992 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-2wbp7_0dc0e269-4d2b-4e4b-afcc-0547ac908a26/kube-rbac-proxy/0.log" Nov 28 15:56:24 crc kubenswrapper[4805]: I1128 15:56:24.002122 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-2wbp7_0dc0e269-4d2b-4e4b-afcc-0547ac908a26/manager/0.log" Nov 28 15:56:24 crc kubenswrapper[4805]: I1128 15:56:24.098925 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-c6lqn_2dac7bca-c6db-40a9-ac64-b46c7722798b/manager/0.log" Nov 28 15:56:24 crc kubenswrapper[4805]: I1128 15:56:24.122186 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-c6lqn_2dac7bca-c6db-40a9-ac64-b46c7722798b/kube-rbac-proxy/0.log" Nov 28 15:56:24 crc kubenswrapper[4805]: I1128 15:56:24.231129 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw_5cb86467-0536-46c2-a399-53bf3ea553b3/kube-rbac-proxy/0.log" Nov 28 15:56:24 crc kubenswrapper[4805]: I1128 15:56:24.263349 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6c5cc85f67hjwcw_5cb86467-0536-46c2-a399-53bf3ea553b3/manager/0.log" Nov 28 15:56:24 crc kubenswrapper[4805]: I1128 15:56:24.581955 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-848c47db98-7l2qm_af063d23-6938-4b22-b491-60e327e03f2b/operator/0.log" Nov 28 15:56:24 crc kubenswrapper[4805]: I1128 15:56:24.697909 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-lz8tb_e163631e-4442-4a96-a1e2-7b0ea21fd517/registry-server/0.log" Nov 28 15:56:24 crc kubenswrapper[4805]: I1128 15:56:24.816395 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-9q29v_8e874a32-dff0-48ad-b43b-a24c364e5e8d/kube-rbac-proxy/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.006969 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-9q29v_8e874a32-dff0-48ad-b43b-a24c364e5e8d/manager/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.039453 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-hnhwx_abe59027-7c48-4b8e-8255-067032ba364d/kube-rbac-proxy/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.183260 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6f5f885fb5-4vb9n_e8c06141-6007-491e-bdd0-6a7654607554/manager/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.253310 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-hnhwx_abe59027-7c48-4b8e-8255-067032ba364d/manager/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.303535 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-m9thn_25d77f9e-dbcf-42b8-8e8e-28d3a500791d/operator/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.392146 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-mg6cr_5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0/kube-rbac-proxy/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.472505 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-mg6cr_5a4bcea2-6232-49b3-89c9-09fbcb1ae5a0/manager/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.505288 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-2jrp8_01f64db7-5c9e-41ea-a595-75dd81511543/kube-rbac-proxy/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.623945 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-2jrp8_01f64db7-5c9e-41ea-a595-75dd81511543/manager/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.657553 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-pmlh7_dc3a6e21-e0ca-40d0-b140-21215085f97a/kube-rbac-proxy/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.809138 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-pmlh7_dc3a6e21-e0ca-40d0-b140-21215085f97a/manager/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.939223 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-2pnsl_519f14b3-c5a1-4bed-8edc-92a84bf0a70d/kube-rbac-proxy/0.log" Nov 28 15:56:25 crc kubenswrapper[4805]: I1128 15:56:25.954441 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-2pnsl_519f14b3-c5a1-4bed-8edc-92a84bf0a70d/manager/0.log" Nov 28 15:56:42 crc kubenswrapper[4805]: I1128 15:56:42.797802 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-lqdbd_4f10987d-9552-4178-98b0-496694b72649/control-plane-machine-set-operator/0.log" Nov 28 15:56:42 crc kubenswrapper[4805]: I1128 15:56:42.934295 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vjvw8_9e9a23de-9ba3-4581-aa91-dfb253372643/kube-rbac-proxy/0.log" Nov 28 15:56:42 crc kubenswrapper[4805]: I1128 15:56:42.949742 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vjvw8_9e9a23de-9ba3-4581-aa91-dfb253372643/machine-api-operator/0.log" Nov 28 15:56:54 crc kubenswrapper[4805]: I1128 15:56:54.699714 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-65t7k_1228d1c5-60b6-4b46-8f2f-f163ff0dea96/cert-manager-controller/0.log" Nov 28 15:56:54 crc kubenswrapper[4805]: I1128 15:56:54.813313 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-g55jh_95ead763-33ad-4a58-905d-7bc10997e6e6/cert-manager-cainjector/0.log" Nov 28 15:56:54 crc kubenswrapper[4805]: I1128 15:56:54.861528 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-rqb4d_8a45f875-20b4-409a-b77a-22305360aa93/cert-manager-webhook/0.log" Nov 28 15:57:06 crc kubenswrapper[4805]: I1128 15:57:06.797646 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-bzkzp_5e74c9e2-3720-438a-a6df-1e9083608caa/nmstate-console-plugin/0.log" Nov 28 15:57:06 crc kubenswrapper[4805]: I1128 15:57:06.974880 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-s4w9z_f62593c7-47a5-40e2-8660-591e3cbe2ad1/nmstate-handler/0.log" Nov 28 15:57:07 crc kubenswrapper[4805]: I1128 15:57:07.039976 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-46m8k_d83b5a88-f387-4751-b525-b1cc7f849322/nmstate-metrics/0.log" Nov 28 15:57:07 crc kubenswrapper[4805]: I1128 15:57:07.053643 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-46m8k_d83b5a88-f387-4751-b525-b1cc7f849322/kube-rbac-proxy/0.log" Nov 28 15:57:07 crc kubenswrapper[4805]: I1128 15:57:07.187216 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-j6g47_82867a57-82fc-4a21-88f5-86636310e157/nmstate-operator/0.log" Nov 28 15:57:07 crc kubenswrapper[4805]: I1128 15:57:07.235515 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-6hn77_82d7ddd8-2fc7-4e21-a235-2559df91b09c/nmstate-webhook/0.log" Nov 28 15:57:21 crc kubenswrapper[4805]: I1128 15:57:21.333200 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-lkhkl_4265aa27-789b-4c6c-9d99-e061e507aea2/kube-rbac-proxy/0.log" Nov 28 15:57:21 crc kubenswrapper[4805]: I1128 15:57:21.523580 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-frr-files/0.log" Nov 28 15:57:21 crc kubenswrapper[4805]: I1128 15:57:21.687329 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-lkhkl_4265aa27-789b-4c6c-9d99-e061e507aea2/controller/0.log" Nov 28 15:57:21 crc kubenswrapper[4805]: I1128 15:57:21.788003 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-reloader/0.log" Nov 28 15:57:21 crc kubenswrapper[4805]: I1128 15:57:21.789303 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-frr-files/0.log" Nov 28 15:57:21 crc kubenswrapper[4805]: I1128 15:57:21.815770 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-metrics/0.log" Nov 28 15:57:21 crc kubenswrapper[4805]: I1128 15:57:21.844834 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-reloader/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.050688 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-reloader/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.062292 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-metrics/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.074509 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-frr-files/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.090003 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-metrics/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.247447 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-metrics/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.253841 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-reloader/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.285843 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/controller/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.297254 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/cp-frr-files/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.438234 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/frr-metrics/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.489941 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/kube-rbac-proxy/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.532993 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/kube-rbac-proxy-frr/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.837907 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/reloader/0.log" Nov 28 15:57:22 crc kubenswrapper[4805]: I1128 15:57:22.913102 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-znh85_0f9d1f71-55cc-4bc7-827f-a1b2c12da36a/frr-k8s-webhook-server/0.log" Nov 28 15:57:23 crc kubenswrapper[4805]: I1128 15:57:23.098071 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8bcbb765d-vgkf5_33947506-e060-4ec4-b7b1-dd52bdcbe7e7/manager/0.log" Nov 28 15:57:23 crc kubenswrapper[4805]: I1128 15:57:23.249938 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5d55dc568c-ln8sb_91105ea4-a3b0-41d6-be1e-73bf1d456a2c/webhook-server/0.log" Nov 28 15:57:23 crc kubenswrapper[4805]: I1128 15:57:23.352184 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l9qgw_8aa6b614-d8e7-4956-afc5-3b30a530b76f/frr/0.log" Nov 28 15:57:23 crc kubenswrapper[4805]: I1128 15:57:23.379563 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6wzdr_85167fd2-c5cc-4c8e-9106-5912aa9d4dee/kube-rbac-proxy/0.log" Nov 28 15:57:23 crc kubenswrapper[4805]: I1128 15:57:23.693016 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6wzdr_85167fd2-c5cc-4c8e-9106-5912aa9d4dee/speaker/0.log" Nov 28 15:57:36 crc kubenswrapper[4805]: I1128 15:57:36.223627 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz_4c270e31-7a0d-4985-b559-6d4ebc5440a2/util/0.log" Nov 28 15:57:36 crc kubenswrapper[4805]: I1128 15:57:36.459679 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz_4c270e31-7a0d-4985-b559-6d4ebc5440a2/util/0.log" Nov 28 15:57:36 crc kubenswrapper[4805]: I1128 15:57:36.499561 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz_4c270e31-7a0d-4985-b559-6d4ebc5440a2/pull/0.log" Nov 28 15:57:36 crc kubenswrapper[4805]: I1128 15:57:36.518223 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz_4c270e31-7a0d-4985-b559-6d4ebc5440a2/pull/0.log" Nov 28 15:57:36 crc kubenswrapper[4805]: I1128 15:57:36.657127 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz_4c270e31-7a0d-4985-b559-6d4ebc5440a2/util/0.log" Nov 28 15:57:36 crc kubenswrapper[4805]: I1128 15:57:36.697635 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz_4c270e31-7a0d-4985-b559-6d4ebc5440a2/extract/0.log" Nov 28 15:57:36 crc kubenswrapper[4805]: I1128 15:57:36.716688 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2xrsz_4c270e31-7a0d-4985-b559-6d4ebc5440a2/pull/0.log" Nov 28 15:57:36 crc kubenswrapper[4805]: I1128 15:57:36.836002 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq_e35268b6-e3cf-4e6b-9248-44904ec877cc/util/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.023977 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq_e35268b6-e3cf-4e6b-9248-44904ec877cc/util/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.039685 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq_e35268b6-e3cf-4e6b-9248-44904ec877cc/pull/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.057812 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq_e35268b6-e3cf-4e6b-9248-44904ec877cc/pull/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.213458 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq_e35268b6-e3cf-4e6b-9248-44904ec877cc/util/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.230297 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq_e35268b6-e3cf-4e6b-9248-44904ec877cc/pull/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.231953 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65whq_e35268b6-e3cf-4e6b-9248-44904ec877cc/extract/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.393127 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95_72ae900f-55af-457e-9785-8d671059454d/util/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.577497 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95_72ae900f-55af-457e-9785-8d671059454d/pull/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.596717 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95_72ae900f-55af-457e-9785-8d671059454d/util/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.607705 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95_72ae900f-55af-457e-9785-8d671059454d/pull/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.736611 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95_72ae900f-55af-457e-9785-8d671059454d/util/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.736950 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95_72ae900f-55af-457e-9785-8d671059454d/pull/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.770553 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bnz95_72ae900f-55af-457e-9785-8d671059454d/extract/0.log" Nov 28 15:57:37 crc kubenswrapper[4805]: I1128 15:57:37.926487 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4d7mg_1e223a05-6a95-4786-a73e-3cfbc00188b9/extract-utilities/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.094211 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4d7mg_1e223a05-6a95-4786-a73e-3cfbc00188b9/extract-content/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.095940 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4d7mg_1e223a05-6a95-4786-a73e-3cfbc00188b9/extract-utilities/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.106092 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4d7mg_1e223a05-6a95-4786-a73e-3cfbc00188b9/extract-content/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.258956 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4d7mg_1e223a05-6a95-4786-a73e-3cfbc00188b9/extract-utilities/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.348219 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4d7mg_1e223a05-6a95-4786-a73e-3cfbc00188b9/extract-content/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.513869 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jpf2f_f98700b4-acf9-4b6c-926f-1a3591a3118c/extract-utilities/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.544435 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4d7mg_1e223a05-6a95-4786-a73e-3cfbc00188b9/registry-server/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.716748 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jpf2f_f98700b4-acf9-4b6c-926f-1a3591a3118c/extract-content/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.718022 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jpf2f_f98700b4-acf9-4b6c-926f-1a3591a3118c/extract-utilities/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.730585 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jpf2f_f98700b4-acf9-4b6c-926f-1a3591a3118c/extract-content/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.873642 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jpf2f_f98700b4-acf9-4b6c-926f-1a3591a3118c/extract-utilities/0.log" Nov 28 15:57:38 crc kubenswrapper[4805]: I1128 15:57:38.886041 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jpf2f_f98700b4-acf9-4b6c-926f-1a3591a3118c/extract-content/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.139008 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ld8lt_82e5203a-95bd-4916-8cec-43da144a6c0a/marketplace-operator/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.141839 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jpf2f_f98700b4-acf9-4b6c-926f-1a3591a3118c/registry-server/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.214052 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4tws_29adeed6-df07-497c-9d0f-6a8b818448c8/extract-utilities/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.305928 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4tws_29adeed6-df07-497c-9d0f-6a8b818448c8/extract-content/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.322937 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4tws_29adeed6-df07-497c-9d0f-6a8b818448c8/extract-utilities/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.354563 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4tws_29adeed6-df07-497c-9d0f-6a8b818448c8/extract-content/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.510511 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4tws_29adeed6-df07-497c-9d0f-6a8b818448c8/extract-content/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.513406 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4tws_29adeed6-df07-497c-9d0f-6a8b818448c8/extract-utilities/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.641365 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4tws_29adeed6-df07-497c-9d0f-6a8b818448c8/registry-server/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.679064 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n77t9_17c1bdf1-aa1e-4857-9708-0c0382cd70fd/extract-utilities/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.838007 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n77t9_17c1bdf1-aa1e-4857-9708-0c0382cd70fd/extract-utilities/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.868941 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n77t9_17c1bdf1-aa1e-4857-9708-0c0382cd70fd/extract-content/0.log" Nov 28 15:57:39 crc kubenswrapper[4805]: I1128 15:57:39.879248 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n77t9_17c1bdf1-aa1e-4857-9708-0c0382cd70fd/extract-content/0.log" Nov 28 15:57:40 crc kubenswrapper[4805]: I1128 15:57:40.021650 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n77t9_17c1bdf1-aa1e-4857-9708-0c0382cd70fd/extract-utilities/0.log" Nov 28 15:57:40 crc kubenswrapper[4805]: I1128 15:57:40.028320 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n77t9_17c1bdf1-aa1e-4857-9708-0c0382cd70fd/extract-content/0.log" Nov 28 15:57:40 crc kubenswrapper[4805]: I1128 15:57:40.338922 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n77t9_17c1bdf1-aa1e-4857-9708-0c0382cd70fd/registry-server/0.log" Nov 28 15:58:41 crc kubenswrapper[4805]: I1128 15:58:41.059764 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:58:41 crc kubenswrapper[4805]: I1128 15:58:41.060494 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:58:43 crc kubenswrapper[4805]: I1128 15:58:43.819093 4805 generic.go:334] "Generic (PLEG): container finished" podID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" containerID="178bdd4a9c60adb2c226f2c50aa447c4f9f71d938bcb0888171001944499a496" exitCode=0 Nov 28 15:58:43 crc kubenswrapper[4805]: I1128 15:58:43.819193 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-stv9f/must-gather-lgp5j" event={"ID":"e9cbde7b-d74e-4257-beb9-3c1557b5029b","Type":"ContainerDied","Data":"178bdd4a9c60adb2c226f2c50aa447c4f9f71d938bcb0888171001944499a496"} Nov 28 15:58:43 crc kubenswrapper[4805]: I1128 15:58:43.820704 4805 scope.go:117] "RemoveContainer" containerID="178bdd4a9c60adb2c226f2c50aa447c4f9f71d938bcb0888171001944499a496" Nov 28 15:58:44 crc kubenswrapper[4805]: I1128 15:58:44.349782 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-stv9f_must-gather-lgp5j_e9cbde7b-d74e-4257-beb9-3c1557b5029b/gather/0.log" Nov 28 15:58:51 crc kubenswrapper[4805]: I1128 15:58:51.695271 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-stv9f/must-gather-lgp5j"] Nov 28 15:58:51 crc kubenswrapper[4805]: I1128 15:58:51.696066 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-stv9f/must-gather-lgp5j" podUID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" containerName="copy" containerID="cri-o://5a3f74e6b9a054b6fde8b003540ccf082ffdabe0f44e766655ee750442c24b12" gracePeriod=2 Nov 28 15:58:51 crc kubenswrapper[4805]: I1128 15:58:51.703397 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-stv9f/must-gather-lgp5j"] Nov 28 15:58:51 crc kubenswrapper[4805]: I1128 15:58:51.907818 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-stv9f_must-gather-lgp5j_e9cbde7b-d74e-4257-beb9-3c1557b5029b/copy/0.log" Nov 28 15:58:51 crc kubenswrapper[4805]: I1128 15:58:51.908292 4805 generic.go:334] "Generic (PLEG): container finished" podID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" containerID="5a3f74e6b9a054b6fde8b003540ccf082ffdabe0f44e766655ee750442c24b12" exitCode=143 Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.101903 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-stv9f_must-gather-lgp5j_e9cbde7b-d74e-4257-beb9-3c1557b5029b/copy/0.log" Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.102803 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.175761 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9cbde7b-d74e-4257-beb9-3c1557b5029b-must-gather-output\") pod \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\" (UID: \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\") " Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.175981 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8w8z\" (UniqueName: \"kubernetes.io/projected/e9cbde7b-d74e-4257-beb9-3c1557b5029b-kube-api-access-m8w8z\") pod \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\" (UID: \"e9cbde7b-d74e-4257-beb9-3c1557b5029b\") " Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.182373 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9cbde7b-d74e-4257-beb9-3c1557b5029b-kube-api-access-m8w8z" (OuterVolumeSpecName: "kube-api-access-m8w8z") pod "e9cbde7b-d74e-4257-beb9-3c1557b5029b" (UID: "e9cbde7b-d74e-4257-beb9-3c1557b5029b"). InnerVolumeSpecName "kube-api-access-m8w8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.264128 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9cbde7b-d74e-4257-beb9-3c1557b5029b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e9cbde7b-d74e-4257-beb9-3c1557b5029b" (UID: "e9cbde7b-d74e-4257-beb9-3c1557b5029b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.277620 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8w8z\" (UniqueName: \"kubernetes.io/projected/e9cbde7b-d74e-4257-beb9-3c1557b5029b-kube-api-access-m8w8z\") on node \"crc\" DevicePath \"\"" Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.277655 4805 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9cbde7b-d74e-4257-beb9-3c1557b5029b-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.916508 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-stv9f_must-gather-lgp5j_e9cbde7b-d74e-4257-beb9-3c1557b5029b/copy/0.log" Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.918095 4805 scope.go:117] "RemoveContainer" containerID="5a3f74e6b9a054b6fde8b003540ccf082ffdabe0f44e766655ee750442c24b12" Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.918234 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-stv9f/must-gather-lgp5j" Nov 28 15:58:52 crc kubenswrapper[4805]: I1128 15:58:52.949283 4805 scope.go:117] "RemoveContainer" containerID="178bdd4a9c60adb2c226f2c50aa447c4f9f71d938bcb0888171001944499a496" Nov 28 15:58:53 crc kubenswrapper[4805]: I1128 15:58:53.213209 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" path="/var/lib/kubelet/pods/e9cbde7b-d74e-4257-beb9-3c1557b5029b/volumes" Nov 28 15:59:11 crc kubenswrapper[4805]: I1128 15:59:11.060040 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:59:11 crc kubenswrapper[4805]: I1128 15:59:11.060554 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:59:41 crc kubenswrapper[4805]: I1128 15:59:41.060438 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:59:41 crc kubenswrapper[4805]: I1128 15:59:41.061046 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:59:41 crc kubenswrapper[4805]: I1128 15:59:41.061091 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" Nov 28 15:59:41 crc kubenswrapper[4805]: I1128 15:59:41.061772 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"482bf7c90ddabbe0361f94c2e9d217342f00d85b78e5e55882137ac7b4c59d0d"} pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:59:41 crc kubenswrapper[4805]: I1128 15:59:41.061871 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" containerID="cri-o://482bf7c90ddabbe0361f94c2e9d217342f00d85b78e5e55882137ac7b4c59d0d" gracePeriod=600 Nov 28 15:59:41 crc kubenswrapper[4805]: I1128 15:59:41.356885 4805 generic.go:334] "Generic (PLEG): container finished" podID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerID="482bf7c90ddabbe0361f94c2e9d217342f00d85b78e5e55882137ac7b4c59d0d" exitCode=0 Nov 28 15:59:41 crc kubenswrapper[4805]: I1128 15:59:41.356939 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerDied","Data":"482bf7c90ddabbe0361f94c2e9d217342f00d85b78e5e55882137ac7b4c59d0d"} Nov 28 15:59:41 crc kubenswrapper[4805]: I1128 15:59:41.357580 4805 scope.go:117] "RemoveContainer" containerID="c9f830a8bd1d510e9306b16e4f3d555b8adb44c770dae581e995a19201aedbed" Nov 28 15:59:42 crc kubenswrapper[4805]: I1128 15:59:42.369425 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" event={"ID":"73da70d8-a7dc-4fca-9e65-9c0d0d815966","Type":"ContainerStarted","Data":"a778e13d55b1f486f8e9bcb2b0b1f8572682ddc2cd4b6ade018cacec0ee1ca5c"} Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.163083 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4"] Nov 28 16:00:00 crc kubenswrapper[4805]: E1128 16:00:00.164179 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" containerName="copy" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.164197 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" containerName="copy" Nov 28 16:00:00 crc kubenswrapper[4805]: E1128 16:00:00.164210 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" containerName="gather" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.164217 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" containerName="gather" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.164428 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" containerName="gather" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.164460 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9cbde7b-d74e-4257-beb9-3c1557b5029b" containerName="copy" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.165010 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.168159 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.168403 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.193351 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4"] Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.290298 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8add3c4-616a-454c-8d16-ed14b0bef0ee-config-volume\") pod \"collect-profiles-29405760-txjz4\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.290412 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8add3c4-616a-454c-8d16-ed14b0bef0ee-secret-volume\") pod \"collect-profiles-29405760-txjz4\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.290441 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrwkt\" (UniqueName: \"kubernetes.io/projected/f8add3c4-616a-454c-8d16-ed14b0bef0ee-kube-api-access-vrwkt\") pod \"collect-profiles-29405760-txjz4\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.391178 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8add3c4-616a-454c-8d16-ed14b0bef0ee-config-volume\") pod \"collect-profiles-29405760-txjz4\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.391493 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8add3c4-616a-454c-8d16-ed14b0bef0ee-secret-volume\") pod \"collect-profiles-29405760-txjz4\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.391582 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrwkt\" (UniqueName: \"kubernetes.io/projected/f8add3c4-616a-454c-8d16-ed14b0bef0ee-kube-api-access-vrwkt\") pod \"collect-profiles-29405760-txjz4\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.392419 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8add3c4-616a-454c-8d16-ed14b0bef0ee-config-volume\") pod \"collect-profiles-29405760-txjz4\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.406936 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8add3c4-616a-454c-8d16-ed14b0bef0ee-secret-volume\") pod \"collect-profiles-29405760-txjz4\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.419123 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrwkt\" (UniqueName: \"kubernetes.io/projected/f8add3c4-616a-454c-8d16-ed14b0bef0ee-kube-api-access-vrwkt\") pod \"collect-profiles-29405760-txjz4\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.487906 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:00 crc kubenswrapper[4805]: I1128 16:00:00.918480 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4"] Nov 28 16:00:01 crc kubenswrapper[4805]: I1128 16:00:01.560522 4805 generic.go:334] "Generic (PLEG): container finished" podID="f8add3c4-616a-454c-8d16-ed14b0bef0ee" containerID="f316c47c7aeef4af93ce1af7067313cc318629f3fc62766dce5bebd996fe7368" exitCode=0 Nov 28 16:00:01 crc kubenswrapper[4805]: I1128 16:00:01.560576 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" event={"ID":"f8add3c4-616a-454c-8d16-ed14b0bef0ee","Type":"ContainerDied","Data":"f316c47c7aeef4af93ce1af7067313cc318629f3fc62766dce5bebd996fe7368"} Nov 28 16:00:01 crc kubenswrapper[4805]: I1128 16:00:01.560620 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" event={"ID":"f8add3c4-616a-454c-8d16-ed14b0bef0ee","Type":"ContainerStarted","Data":"4d81e8461d434f779229061539dd800540ad1d948b668a7666799bc909806f8b"} Nov 28 16:00:02 crc kubenswrapper[4805]: I1128 16:00:02.864834 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.031300 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrwkt\" (UniqueName: \"kubernetes.io/projected/f8add3c4-616a-454c-8d16-ed14b0bef0ee-kube-api-access-vrwkt\") pod \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.031503 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8add3c4-616a-454c-8d16-ed14b0bef0ee-secret-volume\") pod \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.031641 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8add3c4-616a-454c-8d16-ed14b0bef0ee-config-volume\") pod \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\" (UID: \"f8add3c4-616a-454c-8d16-ed14b0bef0ee\") " Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.032610 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8add3c4-616a-454c-8d16-ed14b0bef0ee-config-volume" (OuterVolumeSpecName: "config-volume") pod "f8add3c4-616a-454c-8d16-ed14b0bef0ee" (UID: "f8add3c4-616a-454c-8d16-ed14b0bef0ee"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.033217 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8add3c4-616a-454c-8d16-ed14b0bef0ee-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.041113 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8add3c4-616a-454c-8d16-ed14b0bef0ee-kube-api-access-vrwkt" (OuterVolumeSpecName: "kube-api-access-vrwkt") pod "f8add3c4-616a-454c-8d16-ed14b0bef0ee" (UID: "f8add3c4-616a-454c-8d16-ed14b0bef0ee"). InnerVolumeSpecName "kube-api-access-vrwkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.041465 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8add3c4-616a-454c-8d16-ed14b0bef0ee-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f8add3c4-616a-454c-8d16-ed14b0bef0ee" (UID: "f8add3c4-616a-454c-8d16-ed14b0bef0ee"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.134553 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrwkt\" (UniqueName: \"kubernetes.io/projected/f8add3c4-616a-454c-8d16-ed14b0bef0ee-kube-api-access-vrwkt\") on node \"crc\" DevicePath \"\"" Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.134600 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8add3c4-616a-454c-8d16-ed14b0bef0ee-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.580055 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" event={"ID":"f8add3c4-616a-454c-8d16-ed14b0bef0ee","Type":"ContainerDied","Data":"4d81e8461d434f779229061539dd800540ad1d948b668a7666799bc909806f8b"} Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.580094 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d81e8461d434f779229061539dd800540ad1d948b668a7666799bc909806f8b" Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.580150 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405760-txjz4" Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.944558 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc"] Nov 28 16:00:03 crc kubenswrapper[4805]: I1128 16:00:03.950416 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405715-bnsqc"] Nov 28 16:00:05 crc kubenswrapper[4805]: I1128 16:00:05.219427 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ddd7ac2-635b-4fba-9765-a81039204b8f" path="/var/lib/kubelet/pods/4ddd7ac2-635b-4fba-9765-a81039204b8f/volumes" Nov 28 16:00:36 crc kubenswrapper[4805]: I1128 16:00:36.719321 4805 scope.go:117] "RemoveContainer" containerID="a40ad42cb4c8682b53fd13bb455f69ec3749f5082fd18e9951e96e984877794c" Nov 28 16:01:41 crc kubenswrapper[4805]: I1128 16:01:41.060798 4805 patch_prober.go:28] interesting pod/machine-config-daemon-9pzmp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 16:01:41 crc kubenswrapper[4805]: I1128 16:01:41.061479 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9pzmp" podUID="73da70d8-a7dc-4fca-9e65-9c0d0d815966" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112343577024455 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112343600017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112337071016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112337072015456 5ustar corecore